Summary
Join TrueML, a mission-driven financial software company, as a Senior Data Engineer II and play a pivotal role in designing, building, and maintaining our cutting-edge data LakeHouse platform. You will leverage open table formats like Apache Iceberg to create scalable, reliable data solutions. Develop and operate robust data pipelines, integrating diverse source systems and implementing efficient data transformations for both batch and streaming data. Analyze business requirements and design data models. Develop and manage a scalable AWS cloud infrastructure. Monitor data workloads for performance and errors. Collaborate with Data Services and Data Science colleagues. This role offers unlimited PTO and medical benefit contributions.
Requirements
- Bachelor's degree in Computer Science, Engineering, or a related technical field (Master's degree is a plus)
- 5+ years of hands-on engineering experience (software or data), with a strong emphasis on 3+ years in data-focused roles
- Experience implementing data lake and data warehousing platforms
- Strong Python and SQL skills applied to data engineering tasks
- Proficiency with the AWS data ecosystem, including services like S3, Glue Catalog, IAM, and Secrets Manager
- Experience with Terraform and Kubernetes
- Track record of successfully building and operationalizing data pipelines
- Experience working with diverse data stores, particularly relational databases
Responsibilities
- Building Data LakeHouse: In the Senior Data Engineer II role, you will design, build, and operate robust data lakehouse solutions utilizing open table formats like Apache Iceberg. Your focus will be on delivering a scalable, reliable data lakehouse with optimized query performance for a wide range of analytical workloads and emerging data applications
- Pipeline and Transformation: Integrate with diverse source systems and construct scalable data pipelines. Implement efficient data transformation logic for both batch and streaming data, accommodating various data formats and structures
- Data Modeling: Analyze business requirements and profile source data to design, develop, and implement robust data models and curated data products that power reporting, analytics, and machine learning applications
- Data Infrastructure : Develop and manage a scalable AWS cloud infrastructure for the data platform, employing Infrastructure as Code (IaC) to reliably support diverse data workloads. Implement CI/CD pipelines for automated, consistent, and scalable infrastructure deployments across all environments, adhering to best practices and company standards
- Monitoring and Maintenance: Monitor data workloads for performance and errors, and troubleshoot issues to maintain high levels of data quality, freshness, and adherence to defined SLAs
- Collaboration: Collaborate closely with Data Services and Data Science colleagues to drive the evolution of our data platform, focusing on delivering solutions that empower data users and satisfy stakeholder needs throughout the organization
Preferred Qualifications
- Experience with Airflow, DBT, and Snowflake
- Certification in relevant technologies or methodologies
- Experience with streaming processing technology, e.g., Flink, Spark Streaming
- Familiarity with Domain-Driven Design principles and event-driven architectures
- Certification in relevant technologies or methodologies
Benefits
- Unlimited PTO
- Medical benefit contributions in congruence with local laws and type of employment agreement
Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.