Staff Data Engineer

Apollo.io Logo

Apollo.io

πŸ“Remote - United States

Summary

Join Apollo.io as a Staff Data Engineer and be responsible for maintaining and operating the data platform supporting machine learning workflows and analytics. You will develop and maintain scalable data pipelines, build integrations, and improve data APIs. The role requires implementing monitoring and data quality processes, writing tests and documentation, and defining data models. Collaboration with various teams to strategize long-term data platform architecture is crucial. Success requires a customer-driven approach, high impact focus, ownership, and team mentorship.

Requirements

  • 8+ years of experience as a data platform engineer or a software engineer in data or big data engineer
  • Experience in data modeling, data warehousing, APIs, and building data pipelines
  • Deep knowledge of databases and data warehousing with an ability to collaborate cross-functionally
  • Bachelor's degree in a quantitative field (Physical/Computer Science, Engineering, Mathematics, or Statistics)

Responsibilities

  • Develop and maintain scalable data pipelines and build new integrations to support continuing increases in data volume and complexity
  • Develop and improve Data APIs used in machine learning / AI product offerings
  • Implement automated monitoring, alerting, self-healing (restartable/graceful failures) features while building the consumption pipelines
  • Implement processes and systems to monitor data quality, ensuring production data is always accurate and available
  • Write unit/integration tests, contribute to the engineering wiki, and document work
  • Define company data models and write jobs to populate data models in our data warehouse
  • Work closely with all business units and engineering teams to develop a strategy for long-term data platform architecture

Preferred Qualifications

  • Experience using the Python data stack
  • Experience deploying and managing data pipelines in the cloud
  • Experience working with technologies like Airflow, Hadoop, FastAPI and Spark
  • Understanding of streaming technologies like Kafka and Spark Streaming

Benefits

  • We invest deeply in your growth, ensuring you have the resources, support, and autonomy to own your role and make a real impact
  • Collaboration is at our coreβ€”we’re all for one , meaning you’ll have a team across departments ready to help you succeed
  • We encourage bold ideas and courageous action , giving you the freedom to experiment, take smart risks, and drive big wins

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.

Similar Remote Jobs