Senior Data Engineer

Adaptiq
Summary
Join Adaptiq, a technology hub supporting R&D teams, as a Senior Data Engineer specializing in Python. You will have end-to-end ownership of scalable machine learning pipelines, enhancing data infrastructure, implementing monitoring and observability, and building data validation pipelines. Collaborate with various teams to deliver high-quality pipelines. This role requires a Bachelor's degree in a related field, 5+ years of experience in data engineering and Python, and expertise in SQL, Airflow, and AWS services. The ideal candidate will also possess experience with MLOps, API development, and data validation tools. Adaptiq offers 20 vacation days, full accounting and legal support, a fully remote work model, and a competitive compensation package.
Requirements
- A Bachelorβs or higher in Computer Science, Software Engineering or a closely related technical field, demonstrating strong analytical and coding skills
- At least 5 years of experience as a data engineer, software engineer, or similar role and using data to drive business results
- At least 5 years of experience with Python, building modular, testable, and production-ready code
- Solid understanding of SQL, including indexing best practices, and hands-on experience working with large-scale data systems (e.g., Spark, Glue, Athena)
- Practical experience with Airflow or similar orchestration frameworks, including designing, scheduling, maintaining, troubleshooting, and optimizing data workflows (DAGs)
- A solid understanding of data engineering principles: ETL/ELT design, data integrity, schema evolution, and performance optimization
- Familiarity with AWS cloud services, including S3, Lambda, Glue, RDS, and API Gateway
Responsibilities
- Design and implement scalable machine learning pipelines with Airflow, enabling efficient parallel execution
- Enhance our data infrastructure by refining database schemas, developing and improving APIs for internal systems, overseeing schema migrations, managing data lifecycles, optimizing query performance, and maintaining large-scale data pipelines
- Implement monitoring and observability, using AWS Athena and QuickSight to track performance, model accuracy, operational KPIs and alerts
- Build and maintain data validation pipelines to ensure incoming data quality and proactively detect anomalies or drift
- Collaborate closely with software architects, DevOps engineers, and product teams to deliver resilient, scalable, production-grade machine learning pipelines
Preferred Qualifications
- Experience with MLOps practices such as CI/CD, model and data versioning, observability, and deployment
- Familiarity with API development frameworks (e.g., FastAPI)
- Knowledge of data validation techniques and tools (e.g., Great Expectations, data drift detection)
- Exposure to AI/ML system design, including pipelines, model evaluation metrics, and production deployment
Benefits
- We provide 20 days of vacation leave per calendar year (plus official national holidays of a country you are based in)
- We provide full accounting and legal support in all countries we operate
- We utilize a fully remote work model with a powerful workstation and co-working space in case you need it
- We offer a highly competitive package with yearly performance and compensation reviews