Data Engineer

Komodo Health
Summary
Join Komodo Health, a company dedicated to reducing the global burden of disease through data-driven insights, as a Data Engineer. You will build and scale the data platform, design robust data pipelines, and implement data best practices. Collaborate with various teams to understand data requirements and develop data processing steps. This role is key to the scalability of data systems across product offerings, including designing and implementing data pipelines. In your first year, you will gain a deep understanding of Komodo Health's data landscape and contribute to significant improvements in data infrastructure and product offerings. You will also ensure non-functional requirements are met, such as cost, developer experience, reliability, and maintainability. Komodo Health offers a hybrid work model with flexibility and a competitive benefits package.
Requirements
- Demonstrated proficiency in designing and developing with distributed data processing platforms like Spark and pipeline orchestration tools like Airflow and/or Temporal
- Experience with modern data warehouses such as Snowflake; Experience with SQL and query design on large, complex datasets
- Solid computer science skills and proficiency in programming languages like Python. Able to leverage industry standard engineering best practices, like design patterns and/or testing
- Capable of quickly building expertise on an as-need basis on a new tech stack
- Experience with product engineering software development, in an agile environment
- Demonstrated track-record of delivering products and features with varying degrees of complexity, and through several iterations of product development
- Understand and design for non-functional concerns such as performance, cost optimization, maintainability and developer experience
- A thirst for knowledge, willingness to learn, and a growth-oriented mindset
- Excellent cross-team communication and collaboration skills
Responsibilities
- Partnering with Engineering team members, Product Managers, and customer-facing teams to understand complex health data use cases and business logic
- Being curious about our data
- Building foundational pieces of our data platform architecture, pipelines, analytics, and services underlying our platform
- Designing and developing reliable data pipelines that transform data at scale, orchestrated jobs via Airflow/Temporal, using SQL and Python in Snowflake
- Contributing to python packages in Github and APIs, using current best practices
Preferred Qualifications
- Experience in building containerized API services to serve both internal and external clients
- Experience enhancing CI/CD build tooling in a containerized environment, from deployment pipelines (Jenkins, etc), infrastructure as code (Terraform, Cloudformation), and configuration management via Docker and Kubernetes
- US health care data experience is not required but it is a strong plus
Benefits
- Medical, dental and vision coverage
- 401k Retirement Plan
- Prepaid legal assistance
- Paid time off for vacation, sickness, holiday, and bereavement
- 100% company-paid life insurance and long-term disability insurance