Senior Data Engineer

Voodoo Logo

Voodoo

πŸ“Remote - France

Summary

Join Voodoo's Engineering & Data team as a Data Engineer and contribute to the development and maintenance of scalable, secure, and high-performance data platforms. You will build and manage data pipelines, architect lakehouse solutions, and lead the design and implementation of data infrastructure components. This role requires extensive experience in data engineering, strong programming skills in Python and Java, and a deep understanding of distributed systems. The position offers the opportunity to mentor experienced data engineers and collaborate with cross-functional teams. This role can be fully remote in any EMEA country and offers best-in-class compensation and other benefits according to the country of residence.

Requirements

  • Extensive experience in data engineering or platform engineering roles
  • Strong programming skills in Python and Java
  • Strong experience with modern data stacks (e.g., Spark, Kafka, DBT, Airflow, Lakehouse)
  • Deep understanding of distributed systems, data architecture, and performance tuning
  • Experience with cloud platforms (AWS, GCP, or Azure) and Infrastructure-as-Code tools (Terraform, CloudFormation, etc.)
  • Solid experience operating data services in Kubernetes, including Helm, resource tuning, and service discovery
  • Strong understanding of data modeling, data governance, and security best practices
  • Knowledge of CI/CD principles and DevOps practices in a data environment
  • Excellent problem-solving, communication, and leadership skills

Responsibilities

  • Design, develop, and maintain scalable, secure, and high-performance data platforms
  • Build and manage data pipelines (ETL/ELT) using tools such as Apache Airflow, DBT, SQLMesh or similar
  • Architect and optimize lakehouse solutions (e.g., Iceberg)
  • Lead the design and implementation of data infrastructure components (streaming, batch processing, orchestration, lineage, observability)
  • Ensure data quality, governance, and compliance (GDPR, HIPAA, etc.) across all data processes
  • Automate infrastructure provisioning and CI/CD pipelines for data platform components using tools like Terraform, CircleCI, or similar
  • Collaborate cross-functionally with data scientists, analytics teams, and product engineers to understand data needs and deliver scalable solutions
  • Mentor experienced data engineers and set best practices for code quality, testing, and platform reliability
  • Monitor and troubleshoot performance issues in real-time data flows and long-running batch jobs
  • Stay ahead of trends in data engineering, proactively recommending new technologies and approaches to keep our stack modern and efficient

Preferred Qualifications

  • Experience with real-time data streaming and event-driven architectures
  • Familiarity with ML model deployment and MLOps practices
  • Exposure to data cataloging, lineage tools and observability platforms
  • Contributions to open-source data tools or platforms

Benefits

  • Best-in-class compensation
  • Other benefits according to the country you reside in

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.