It starts with you – an engineer who cares about building data pipelines and models that deliver reliable, trusted data. You value data quality, clean transformations, and making data accessible to those who need it. Youll work alongside experienced engineers to build ETL/ELT pipelines, maintain dimensional models, and implement quality checks that turn raw data into actionable intelligence.
If you want to grow your skills building data products for mission-critical AI, join mission – this role is for you.
:Responsibilities
Build and maintain ETL/ELT pipelines using platform tooling – workflows that extract from sources, apply transformations, and load into analytical stores.
Develop and maintain data models – fact/dimension tables, aggregations, and views that serve analytics and ML use cases.
Implement data quality checks – validation rules, tests, and monitoring for data freshness and accuracy.
Maintain documentation and lineage – keeping data catalogs current and helping consumers understand data sources and transformations.
Work with stakeholders to understand data requirements and implement requested data products.
Troubleshoot pipeline failures – investigating issues, fixing bugs, and improving reliability.
Write clean, tested, well-documented SQL and Python code.
Collaborate with Data Platform on tooling needs; work with Datastores on database requirements; partner with ML, Data Science, Analytics, Engineering, and Product teams on data needs.
Design retrieval-friendly data artifacts – RAG-supporting views, feature tables, and embedding pipelines – with attention to freshness and governance expectations.
If you want to grow your skills building data products for mission-critical AI, join mission – this role is for you.
:Responsibilities
Build and maintain ETL/ELT pipelines using platform tooling – workflows that extract from sources, apply transformations, and load into analytical stores.
Develop and maintain data models – fact/dimension tables, aggregations, and views that serve analytics and ML use cases.
Implement data quality checks – validation rules, tests, and monitoring for data freshness and accuracy.
Maintain documentation and lineage – keeping data catalogs current and helping consumers understand data sources and transformations.
Work with stakeholders to understand data requirements and implement requested data products.
Troubleshoot pipeline failures – investigating issues, fixing bugs, and improving reliability.
Write clean, tested, well-documented SQL and Python code.
Collaborate with Data Platform on tooling needs; work with Datastores on database requirements; partner with ML, Data Science, Analytics, Engineering, and Product teams on data needs.
Design retrieval-friendly data artifacts – RAG-supporting views, feature tables, and embedding pipelines – with attention to freshness and governance expectations.
Requirements:
3+ years in data engineering, analytics engineering, BI development, or software engineering with strong SQL focus.
Strong SQL skills; complex queries, window functions, CTEs, query optimization basics
Data modeling – Understanding of dimensional modeling concepts; fact/dimension tables, star schemas
Transformation frameworks – Exposure to dbt, Spark SQL, or similar; understanding of modular, testable transformations
Orchestration – Familiarity with Airflow, Dagster, or similar; understanding of DAGs, scheduling, dependencies
Data quality – Awareness of data validation approaches, testing strategies, and quality monitoring
Python – Proficiency in Python for data manipulation and scripting; pandas, basic testing
Version control – Git workflows, code review practices, documentation
3+ years in data engineering, analytics engineering, BI development, or software engineering with strong SQL focus.
Strong SQL skills; complex queries, window functions, CTEs, query optimization basics
Data modeling – Understanding of dimensional modeling concepts; fact/dimension tables, star schemas
Transformation frameworks – Exposure to dbt, Spark SQL, or similar; understanding of modular, testable transformations
Orchestration – Familiarity with Airflow, Dagster, or similar; understanding of DAGs, scheduling, dependencies
Data quality – Awareness of data validation approaches, testing strategies, and quality monitoring
Python – Proficiency in Python for data manipulation and scripting; pandas, basic testing
Version control – Git workflows, code review practices, documentation
This position is open to all candidates.


















