A leading provider of technology-driven data solutions that empower businesses with transformative insights, this company is a pioneer in data intelligence and is currently seeking to build a new, predictive model.
Check all associated application documentation thoroughly before clicking on the apply button at the bottom of this description.
In this role, you will design, build, and maintain high-performance data pipelines using tools like Dagster and Python. You'll process diverse data sources, collaborate with data teams, and optimize data workflows for both real-time and batch processing. You'll also contribute to data architecture, ensure data governance and security, and maintain comprehensive documentation.
Technical Skills:
* Proficiency in pipeline orchestration tools (e.g., Dagster, Airflow)
* Strong Python programming skills with experience in libraries like Pandas and PySpark
* Expertise in cloud platforms, particularly AWS (e.g., S3, Lambda, Redshift, RDS)
* Deep understanding of data modeling, ETL workflows, and scalable architecture design
* Familiarity with integrating machine learning models into production workflows
Experience:
* 3+ years of experience in a Data Engineering role, with experience in mid-to-senior capacity
* Proven track record of working with live datasets and building end-to-end data pipelines
* Strong SQL skills for querying and managing large datasets
* Hands-on experience in creating architectural diagrams and delivering technical presentations
A big bonus if you have..
* Experience in building and optimizing propensity models or similar predictive analytics models
* Strong understanding of feature engineering, predictive modeling, and evaluation metrics
* Experience with version control tools (e.g., Git) and CI/CD pipelines
Intrigued? Drop me a message!