Data Architect

Able Logo

Able

πŸ“Remote

Summary

Join Able as a Data Engineer and contribute to the development of robust and scalable data solutions for our clients. Working closely with the Director of Engineering and cross-functional teams, you will lead strategic architecture, build ETL/ELT pipelines, and optimize data systems. This role requires 10+ years of experience in data engineering with expertise in Apache Spark and Delta Lake. You will collaborate with clients, translate business needs into technical solutions, and ensure data quality and compliance. The position is a 4-6 month contract, fully remote within LatAm, requiring strong English communication skills. A contract extension is possible based on performance and client needs.

Requirements

  • 10+ years of data engineering experience with enterprise-scale systems
  • Expertise in Apache Spark and Delta Lake, including ACID transactions, time travel, Z-ordering, and compaction
  • Deep knowledge of Databricks (Jobs, Clusters, Workspaces, Delta Live Tables, Unity Catalog)
  • Experience building scalable ETL/ELT pipelines using tools like Airflow, Glue, Dataflow, or ADF
  • Advanced SQL for data modeling and transformation
  • Strong programming skills in Python (or Scala)
  • Hands-on experience with data formats such as Parquet, Avro, and JSON
  • Familiarity with schema evolution, versioning, and backfilling strategies
  • Working knowledge of at least one major cloud platform: AWS (S3, Athena, Redshift, Glue Catalog, Step Functions) GCP (BigQuery, Cloud Storage, Dataflow, Pub/Sub) – nice to have Azure (Synapse, Data Factory, Azure Databricks) – nice to have
  • Experience designing data architectures with real-time or streaming data (Kafka, Kinesis)
  • Consulting or client-facing experience with strong communication and leadership skills
  • Experience with data mesh architectures and domain-driven data design
  • Knowledge of metadata management, data cataloging, and lineage tracking tools
  • Awareness of international data privacy regulations and compliant system design
  • Strong verbal and written communication skills in English are a requirement

Responsibilities

  • Shape large-scale data architecture vision and roadmap across client engagements
  • Establish governance, security frameworks, and regulatory compliance standards
  • Lead strategy around platform selection, integration, and scaling
  • Guide organizations in adopting data lakehouse and federated data models
  • Lead technical discovery sessions to understand client needs
  • Translate complex architectures into clear, actionable value for stakeholders
  • Build trusted advisor relationships and guide strategic decisions
  • Align architecture recommendations with business growth and goals
  • Design and implement modern data lakehouse architectures with Delta Lake and Databricks
  • Build and manage ETL/ELT pipelines at scale using Spark (PySpark preferred)
  • Leverage Delta Live Tables, Unity Catalog, and schema evolution features
  • Optimize storage and queries on cloud object storage (e.g., AWS S3, Azure Data Lake)
  • Integrate with cloud-native services like AWS Glue, GCP Dataflow, and Azure Synapse Analytics
  • Implement data quality monitoring, lineage tracking, and schema versioning
  • Build scalable pipelines with tools like Apache Airflow, Step Functions, and Cloud Composer
  • Develop cost-optimized, scalable, and compliant data solutions
  • Design POCs and pilots to validate technical approaches
  • Translate business requirements into production-ready data systems
  • Define and track success metrics for platform and pipeline initiatives

Preferred Qualifications

  • Master's degree in Computer Science, Data Engineering, or related field
  • ML Ops experience or integrating machine learning models into data pipelines
  • Relevant certifications in cloud platforms or data engineering
  • Familiarity with healthcare standards (e.g., HL7, FHIR, DICOM) is a plus

Benefits

  • This is a contract position
  • This position is 100% remote within LatAm
  • This contract period is for 4-6 months, starting in August 2025
  • Candidates are expected to work 40 hours per week during this contract period, and be available during normal business hours as-needed on this project
  • A contract extension is possible, pending our client partnership and individual performance

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.