Remote Senior Data Engineer

Logo of Workstream

Workstream

πŸ“Remote - China

Job highlights

Summary

Join Workstream as a seasoned Data Engineer and contribute to building premium software solutions for hourly businesses. In this unique role, you will focus on data engineering tasks while leveraging your ML expertise to collaborate with data scientists for model deployment.

Requirements

  • Bachelor's/Master's degree in Computer Science, Data Science, or a related quantitative field
  • Proficiency in Python and software engineering
  • Proven experience as a Data Engineer, with a solid understanding of SQL, and Big Data technologies
  • Expertise in containerization and orchestration technologies like Docker and Kubernetes
  • Knowledge of vector stores, databases, and data warehousing concepts
  • Experience in deploying, and monitoring ML API services using Flask or FastAPI
  • Strong project management skills, with the ability to collaborate effectively with cross-functional teams

Responsibilities

  • Design, build, and maintain efficient, reliable, and complex ETL pipelines to process and analyze large volumes of data from various sources
  • Develop and enhance our data lakehouse, driving data quality across departments and building self-service tools for analysts
  • Define, build, and own data architecture for a trusted, governed, dimensionally-modeled repository of data
  • Collaborate with cross-functional teams including data scientists to assist in deploying and monitoring machine learning models in production environments
  • Help data scientists develop and maintain ML API services for seamless integration into the company's infrastructure
  • Apply knowledge of real-time, streaming, and batch processing concepts to optimize model performance and data handling
  • Participate in code and design reviews to maintain high development standards

Preferred Qualifications

  • Experience with Hevo Data or other streaming vendors(Fivetran, Airbyte, DMS)
  • Experience with DBT
  • Experience with Snowflake or Redshift
  • Experience with orchestration tools such as Airflow
  • Experience with data catalog solutions such as Atlan
  • Experience with Metaflow is a plus
  • Experience with cloud platforms such as AWS, GCP, or Azure
  • Experience with specialized ML serving tools like Bento, Seldon Core, Hugging Face Inference, Sagemaker Endpoints is a plus

Benefits

  • Competitive salary and equity
  • Comprehensive health coverage
  • Performance-based year-end bonuses
  • Unlimited PTO
  • Remote/WFH schedule

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.
Please let Workstream know you found this job on JobsCollider. Thanks! πŸ™