Senior Data Engineer

Komodo Health Logo

Komodo Health

πŸ“Remote - United States

Summary

Join Komodo Health, a company dedicated to reducing the global burden of disease through data-driven insights, as a Data Engineer. You will play a pivotal role in building and enhancing our data platform, collaborating with cross-functional teams to solve complex data challenges. Responsibilities include designing and implementing data processing and transformation at scale, building robust data pipelines, and contributing to Python packages and APIs. You will leverage your expertise in SQL, Python, and distributed data processing platforms like Spark and Airflow. The ideal candidate possesses a strong understanding of data modeling, schema design, and data storage best practices. Komodo Health offers a hybrid work model and a competitive benefits package.

Requirements

  • Expertise in writing enterprise-level code and contributing to large data pipelining and API processing with Python
  • Experience with SQL and query design on large, complex datasets
  • Ability to use a variety of relational, NoSQL, Postgres, and/or MPP databases (ideally Snowflake on AWS) and leading data modeling, schema design, and data storage best practices
  • Demonstrated proficiency in designing and developing with distributed data processing platforms like Spark and pipeline orchestration tools like Airflow
  • A thirst for knowledge, willingness to learn, and a growth-oriented mindset
  • Committed to fostering an inclusive environment where your teammates feel motivated to succeed
  • Excellent cross-team communication and collaboration skills
  • Ability to leverage AI tools (Gemini, ChatGPT, Cursor, etc.) to enhance personal productivity, streamline workflows, or improve decision-making

Responsibilities

  • Partnering with Engineering team members, Product Managers, Data Scientists, and customer-facing teams to understand complex health data use cases and business logic
  • Being curious about our data
  • Building foundational pieces of our data platform architecture, pipelines, analytics, and services underlying our platform
  • Architecting and developing reliable data pipelines that transform data at scale, orchestrated jobs via Airflow, using SQL and Python in Snowflake
  • Contributing to python packages in Github and APIs, using current best practices

Preferred Qualifications

  • Experience enhancing CI/CD build tooling in a containerized environment, from deployment pipelines (Jenkins, etc), infrastructure as code (Terraform, Cloudformation), and configuration management via Docker and Kubernetes
  • US health care data experience is not required but it is a strong plus
  • Experience interacting with, managing, or building with AI agentic workflows

Benefits

  • Medical, dental and vision coverage
  • 401k Retirement Plan
  • Prepaid legal assistance
  • Paid time off for vacation, sickness, holiday, and bereavement
  • 100% company-paid life insurance and long-term disability insurance

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.

Similar Remote Jobs