Data Engineer

CDC Foundation
Summary
Join the CDC Foundation as a Data Engineer and play a crucial role in advancing its mission by designing, building, and maintaining data infrastructure for a public health organization. This grant-funded, limited-term position (until June 30, 2026) is part of the Workforce Acceleration Initiative (WAI), focused on improving public health agencies' information systems. Working remotely from anywhere in the US, you will collaborate with various teams to develop and deploy a mobile data entry/collection mechanism for the Louisville Metro Department of Public Health and Wellness. The role involves creating and managing data pipelines, optimizing infrastructure, implementing security measures, and collaborating with stakeholders. The salary range is $103,500-$143,500 per year, plus benefits, with office hours requiring availability between 8:00 am โ 5:00 pm Eastern Time.
Requirements
- Bachelorโs degree in Computer Science, Information Technology, Data Science, or a related field. Bachelorโs degree in business administration, Engineering Technology, Liberal Arts, Physical Science, Public Administration, Social Science, or related discipline will be considered
- Minimum of 5 years of relevant professional experience. Six years of relational database management system (i.e. human resource management system, MS SQL) experience in a client/server environment preferred
- Proficiency in programming languages commonly used in data engineering, such as Python, Java, SQL; including C#, C++
- Candidate should be able to implement data automations within existing frameworks as opposed to writing one off scripts
- Strong understanding of relational database systems (e.g., Microsoft SQL, MySQL, PostgreSQL)
- Experience with AWS tools for data engineering
- Experience regarding engineering best practices such as source control, automated testing, continuous integration and deployment, and peer review
- Knowledge of data warehousing concepts and tools
- Knowledgeable about industry trends, best practices, and emerging technologies in data engineering, and incorporating the trends into the organization's data infrastructure
- Experience with cloud computing platforms
- Expertise in data modeling, ETL (Extract, Transform, Load) processes, and data integration techniques
- Familiarity with agile development methodologies, software design patterns, and best practices
- Strong analytical thinking and problem-solving abilities
- Excellent verbal and written communication skills, including the ability to convey technical concepts to non-technical partners effectively
- Flexibility to adapt to evolving project requirements and priorities
- Outstanding interpersonal and teamwork skills; and the ability to develop productive working relationships with colleagues and partners
- Experience working in a virtual environment with remote partners and teams
Responsibilities
- Create and manage the systems and pipelines that enable efficient and reliable flow of data, including ingestion, processing, and storage
- Collect data from various sources, transforming and cleaning it to ensure accuracy and consistency. Load data into storage systems or data warehouses
- Optimize data pipelines, infrastructure, and workflows for performance and scalability
- Monitor data pipelines and systems for performance issues, errors, and anomalies, and implement solutions to address them
- Implement security measures to protect sensitive information
- Collaborate with data engineers, analysts, and other partners to understand their data needs and requirements, and to ensure that the data infrastructure supports the organization's goals and objectives
- Collaborate with cross-functional teams to understand data requirements and design scalable solutions that meet business needs
- Implement and maintain ETL processes to ensure the accuracy, completeness, and consistency of data
- Design and manage data storage systems, including relational databases
- Provide technical guidance to other staff
- Communicate effectively with partners at all levels of the organization to gather requirements, provide updates, and present findings
Preferred Qualifications
Experience with big data technologies and frameworks like Hadoop, Spark, Kafka, and Flink a plus
Benefits
- Salary Range: $103,500-$143,500 per year, plus benefits
- Position Type: Grant funded, limited-term opportunity
- Position End Date: June 30, 2026
- Location: Remote, must be based in the United States
- Office Hours: Availability for meetings between 8:00 am โ 5:00 pm Eastern Time with flexibility dependent on employee location