Staff Data Engineer
Correlation One
Summary
Join Correlation One as a Staff Data Engineer and become our data lake subject matter expert. You will own the long-term technical vision for our data engineering architecture, developing data products and platforms for flexible, scalable, and reliable data pipelines. Collaborate with internal architects, create reusable frameworks, mentor junior engineers, and monitor database performance. This remote role, based in Latin America and compatible with EST, offers competitive compensation, unlimited time off, company-paid holidays, access to free data skills training, and a supportive company culture. The ideal candidate possesses 7+ years of data engineering experience, advanced SQL skills, and expertise in cloud-based data lake environments. Experience with ETL/ELT processes, microservices, and agile methodologies is also required.
Requirements
- 7+ years in a Data Engineering role with experience in data warehouses, fetching data from APIs, developing and maintaining ETL/ELT to service data marts, software applications or cloud microservices used to support data science applications
- Experience with creating data products and internal platforms which accelerate the development of data pipelines
- Advanced SQL experience and database design
- Experience building and maintaining pipeline monitoring for latency, traffic, saturation, and errors
- Able to demonstrate advanced computer and analytical skills with particular knowledge and understanding of the following storage, computing, and tools: GCP or equivalent cloud-based data lake/OLAP/OLTP environments
- APIs, Airflow, Cron, dbt, Git
- Experience creating microservices for passing data
- Experience working and implementing CDC (PostGres), with technologies like Kafka
- Working experience with Software Engineering development and deployment practices
- Experience with agile methodologies such as Scrum or Kanban with comfort level to work in iterative product driven cycles
- Experience with object-oriented/object function scripting languages: Python, Scala, etc
- Working with CI/CD processes and source control tools such as Github
Responsibilities
- Own and develop the long-term technical vision & blueprint for the team
- Design technical strategy to develop well-architected data lakehouse
- Collaborate with internal architects to design and build out the ELT process from data ingestion to analytics marts
- Create reliable, reusable frameworks and abstractions to standardize software development
- Provide support to development teams while mentoring engineering team members on modern database architecture principles and best practices
- Monitor and analyze database performance and follow best practices of data engineering such as code reviews, scrum, and SDLC
- Develop schema design for reports and analytics
- Hands-on development from microservices and sub-systems to the entire technical stack
- Deal with ambiguity and figure things out with minimal guidance
- Mentor junior engineers on the team
- Create standards for engineering and operational excellence
- Identify and educate on industry relevant technical trends
Preferred Qualifications
- Handling unstructured text and data
- Infrastructure as code, ideally Terraform
- Non-tabular database management systems, e.g. MongoDB, NeptuneDB
- BI dashboards such as Tableau or PowerBI
- Implementing machine learning systems
- Utilizing logging/observability software such as data dog
- Utilizing data cataloging tool to document data pipelines and datasets
Benefits
- Insurance or subsidies based on country
- Unlimited Time Off, with a minimum time off recommendation
- 10 company-paid holidays
- Official company-wide holiday for the last week of the calendar year
- Access to free data skills training through our programs
- A company culture that empowers individuals and embraces diversity through its core mission