Summary
Join Route's Engineering team as a Staff Data Engineer and play a pivotal role in maintaining and expanding our data infrastructure. This highly autonomous role requires a seasoned engineer with strong technical skills to ensure the continuous and reliable ingestion of data. You will design and implement new data integrations, work cross-functionally with various teams, and leverage technologies like Snowflake, Databricks, and AWS. The position demands expertise in data processing, pipeline development, and cloud infrastructure. Route offers a comprehensive benefits package including 100% health insurance premium coverage, remote work options, unlimited PTO, 401k matching, and professional development opportunities.
Requirements
- 5+ years of experience in data engineering or a similar field
- Expertise in Snowflake, DBT, Databricks, and Spark for data processing and analytics
- Strong programming skills in Python for data pipeline development and automation
- Experience with Snaplogic, Airflow, or similar ETL orchestration tools
- Proficiency in cloud infrastructure and data services within the AWS ecosystem
- Hands-on experience with Terraform for infrastructure as code
- Strong problem-solving skills with the ability to work independently in a fast-paced environment
- Excellent communication skills and ability to collaborate across teams
- Familiarity with real-time data processing architectures
- Knowledge of data governance, security, and compliance best practices
Responsibilities
- Own and maintain the reliability of existing data ingestion pipelines, ensuring uninterrupted data flow
- Design, develop, and optimize new data ingestion and integration pipelines
- Work cross-functionally with data analysts, product managers, back-end and front-end engineering teams, and business stakeholders to deliver robust data solutions
- Leverage Snowflake, Databricks, DBT, and Spark to build scalable and high-performance data processing solutions
- Implement and maintain Snaplogic, Airflow, and Terraform for ETL orchestration, workflow automation, and infrastructure as code
- Ensure high availability, scalability, and security of data infrastructure using the AWS suite
- Troubleshoot and optimize performance across the data pipeline stack
- Stay ahead of industry trends and continuously improve data engineering best practices
- Partner with Data Analysts to enhance analytics and provide guidance on queries that impact multiple departments
Preferred Qualifications
- Experience with AI agents to interact with a database
- Experience in the e-commerce or post-purchase space is a plus
Benefits
- 100% of your health insurance premiums on a $0 deductible plan for you and your family
- Remote or hybrid work arrangements
- Unlimited PTO
- 401k matching
- Formalized growth opportunities
- Learning & development
- DEI programs & events
Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.