Data Engineer

phData Logo

phData

πŸ“Remote - United States

Summary

Join phData, a leading innovator in the modern data stack, partnering with major cloud platforms. We're a remote-first global company committed to helping enterprises solve complex data challenges. We offer a casual, exciting work environment with opportunities for autonomy and growth. We're seeking a highly skilled Data Engineer, Software Engineer, or Solutions Architect with extensive experience in cloud data platforms and data solutions. The ideal candidate will possess strong programming skills, excellent communication abilities, and a proven track record of success. phData is an award-winning workplace that celebrates diversity and inclusion.

Requirements

  • Have 5+ years of experience as a hands-on Data Engineer, Software Engineer, or Solutions Architect designing and implementing data solutions
  • Possess programming expertise in Java, Python, and/or Scala
  • Demonstrate expertise in core cloud data platforms including Snowflake, AWS, Azure, Databricks, and GCP
  • Be proficient in SQL and possess the ability to write, debug, and optimize SQL queries
  • Have client-facing written and verbal communication skills and experience
  • Be capable of creating detailed solution documentation (e.g., including POCs and roadmaps, sequence diagrams, class hierarchies, logical system views, etc.)
  • Hold a 4-year Bachelor's degree in Computer Science or a related field

Responsibilities

Design and implement data solutions

Preferred Qualifications

  • Have production experience in core data platforms: Snowflake, AWS, Azure, GCP, Hadoop, Databricks
  • Have experience with Cloud and Distributed Data Storage: S3, ADLS, HDFS, GCS, Kudu, ElasticSearch/Solr, Cassandra, or other NoSQL storage systems
  • Have experience with Data integration technologies: Spark, Kafka, event/streaming, Streamsets, Matillion, Fivetran, NiFi, AWS Data Migration Services, Azure DataFactory, Informatica Intelligent Cloud Services (IICS), Google DataProc, or other data integration technologies
  • Have experience with Multiple data sources (e.g., queues, relational databases, files, search, API)
  • Have complete software development lifecycle experience including design, documentation, implementation, testing, and deployment
  • Have experience with Automated data transformation and data curation: dbt, Spark, Spark streaming, automated pipelines
  • Have experience with Workflow Management and Orchestration: Airflow, AWS Managed Airflow, Luigi, NiFi

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.