Summary
Join Foxbox Digital, an award-winning digital product agency, as a Data Engineer. You will design, build, and maintain data pipelines and integrations for AI/ML applications. Collaborate with data scientists and MLOps engineers to ensure data quality and reliability. This role requires experience with ETL/ELT pipelines, various data sources, and cloud-based solutions. Foxbox offers a remote-first environment, continuous training, and a collaborative team culture.
Requirements
- Have a Bachelorโs or Masterโs degree in Computer Science, Information Systems, or a related field
- Have 3+ years of hands-on experience in data engineering, building ETL/ELT pipelines, and managing data workflows in a production environment
- Be proficient in Python or Java for data processing
- Be familiar with SQL and NoSQL databases
- Have experience with cloud-based data solutions (AWS, Azure, or GCP)
- Have knowledge of distributed data processing (e.g., Spark, Hadoop, Databricks)
- Excel at working cross-functionally and translating data requirements into actionable solutions
- Have the ability to debug complex data issues quickly and innovate new approaches to data challenges
Responsibilities
- Build and maintain scalable ETL/ELT pipelines, transforming raw data into ready-to-use datasets for machine learning
- Connect to various data sources (e.g., Quickbase, REST APIs, databases) and ensure seamless data ingestion
- Implement validation rules, data profiling, and monitoring to maintain high data integrity
- Work with data scientists and MLOps engineers to optimize data workflows, improve model performance, and address bottlenecks
- Identify and resolve issues around data throughput and latency, ensuring pipelines can handle production-level loads
- Enforce data privacy best practices and role-based access control in accordance with project or regulatory requirements
- Set up alerts, logs, and metrics to proactively detect and resolve pipeline or data-related issues
Preferred Qualifications
- Have experience with Airflow, Kafka, or Azure Data Factory
- Have familiarity with CI/CD practices and Git-based workflows
- Have exposure to MLOps tools like MLflow or Kubeflow
Benefits
- We offer continuous training and growth opportunities
- Remote-first environment with a culture of collaboration and innovation
- Opportunity to work on a project that directly impacts business success
- You are part of a multicultural and collaborative team that is constantly growing
Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.