Data Engineer - Data Operations

Arine Logo

Arine

πŸ’΅ $165k-$180k
πŸ“Remote - United States

Summary

Join Arine, a rapidly growing healthcare technology company, as a Senior Data Engineer and play a pivotal role in building and maintaining our data operations infrastructure. You will architect, build, and maintain scalable data ingestion infrastructure, focusing on the "EL" (Extract-Load) portion of our ELT stack. Collaborate with analytics engineers to ensure seamless data flow into the staging layer for dbt transformations. This role requires expertise in scalable data migration, event-driven pipelines, containerization, and building maintainable toolsets. You will work with large-scale healthcare datasets and contribute to a team committed to improving patient health outcomes. Arine offers a fast-paced environment, a strong team, and the opportunity to make a significant impact on healthcare.

Requirements

  • 6+ years of professional experience in data engineering with focus on large-scale data ingestion and infrastructure
  • Strong experience with scalable data ingestion tools such as Kinesis, Airbyte, Kafka, or similar open-source solutions
  • Proven experience building event-driven ETL/ELT systems that move large datasets from operational databases (RDS, DynamoDB) to data warehouses (Snowflake)
  • Deep understanding of software engineering principles including test-driven development, loose coupling, single responsibility, and modular design
  • Experience with containerization technologies (Docker, Kubernetes) and building configuration-driven, maintainable systems
  • Understanding of medallion/layered data architecture patterns and experience supporting analytics engineering workflows
  • Experience with incremental data processing and change data capture (CDC) methodologies
  • Hands-on experience with cloud data infrastructure, particularly AWS services (S3, Kinesis, Lambda, Step Functions, RDS, DynamoDB)
  • Proven ability to build tools and systems that can be operated by diverse engineering profiles through configuration rather than code changes
  • Experience working with large healthcare datasets and understanding of data privacy and compliance requirements
  • Demonstrated ability to refactor and improve existing data infrastructure for better scalability and operational efficiency
  • Strong collaboration skills working with analytics engineers, data scientists, and ML engineers
  • Excellent verbal and written communication skills with ability to explain technical infrastructure concepts to diverse audiences
  • Passion for building robust, maintainable, and operationally excellent data systems
  • Ability to pass a background check
  • Must live in and be eligible to work in the United States
  • An established private work area that ensures information privacy
  • A stable high-speed internet connection for remote work

Responsibilities

  • Architecting and implementing scalable data ingestion infrastructure from platform sources (RDS, DynamoDB) into Snowflake
  • Building event-driven data pipelines using tools like Kinesis, Airbyte, or other open-source ingestion frameworks that scale effectively
  • Designing systems that support our medallion architecture and enable smooth data flow into the staging layer
  • Creating configuration-driven, containerized toolsets that can be easily used and maintained by diverse engineering profiles
  • Collaborating with analytics engineers to ensure smooth data flow into the staging layer for dbt transformations
  • Implementing incremental data migration strategies for large-scale healthcare datasets
  • Building monitoring and alerting systems for data ingestion processes and pipeline health
  • Applying software engineering best practices including test-driven development and modular design to data infrastructure
  • Refactoring and rebuilding existing data ingestion processes to improve scalability and operational efficiency
  • Working with containerization technologies (Docker, Kubernetes) to create portable and maintainable data solutions
  • Supporting the migration to our staging β†’ intermediate β†’ marts medallion structure
  • Mentoring team members on data operations best practices and infrastructure design patterns

Benefits

$165,000-180,000/year

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.