Data Engineer

Xebia Poland Logo

Xebia Poland

πŸ“Remote - Worldwide

Summary

Join Xebia, a global leader in digital solutions, and become a Data Engineer. You will develop and maintain data pipelines, ensuring data integrity and seamless flows across various systems. Responsibilities include integrating data from diverse sources, implementing ETL processes, and collaborating with cross-functional teams. This role requires 3+ years of experience in data engineering, proficiency in Python and SQL, and expertise with Apache Airflow. A Bachelor's or Master's degree in a related field is also required. The ideal candidate will possess experience with modern data libraries and frameworks. Work from the European Union region and a work permit are required.

Requirements

  • 3+ years in a data engineering role, with hands-on experience in building data processing pipelines
  • Proficiency with Python
  • Proficiency with SQL (large joins, window functions)
  • Extensive experience with Apache Airflow, including DAG creation, triggers, and workflow optimization
  • Knowledge of data partitioning, batch configuration, and performance tuning for terabyte-scale processing
  • Hands-on experience with modern data libraries and frameworks (e.g., Databricks, Snowflake, Spark)
  • Hands-on experience with ETL tools and processes
  • Deep understanding of relational and NoSQL databases, data modelling, and data warehousing concepts
  • Excellent command of oral and written English
  • Available to start within a short time frame (a maximum one month’s notice)
  • Bachelor's or Master’s degree in Computer Science, Information Systems, or a related field
  • Work from the European Union region and a work permit are required

Responsibilities

  • Develop and maintain data pipelines to ensure seamless data flows
  • Ensure data integrity, consistency, and availability across all data systems
  • Integrate data from various sources, including transactional databases, third-party APIs, and external data sources, into the data lake
  • Implement ETL processes to transform and load data into the data warehouse for analytics and reporting
  • Work closely with cross-functional teams including Engineering, Business Analytics, Data Science and Product Management to understand data requirements and deliver solutions
  • Collaborate with data engineers to ensure data engineering best practices are integrated into the development process
  • Optimize data storage and retrieval to improve performance and scalability
  • Monitor and troubleshoot data pipelines to ensure high reliability and efficiency
  • Implement and enforce data governance policies to ensure data security, privacy, and compliance
  • Develop documentation and standards for data processes and procedures

Preferred Qualifications

  • Terraform
  • GitHub Actions
  • AWS/Azure/GCP

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.