Summary
Join phData, a leading innovator in the modern data stack, partnering with major cloud data platforms. We are a remote-first global company committed to helping global enterprises overcome their toughest data challenges. We offer a casual, exciting work environment where top performers have autonomy to deliver results. We're looking for a hands-on Data Engineer with 8+ years of experience designing and implementing data solutions. The ideal candidate will have expertise in Java, Python, or Scala, core cloud data platforms, and SQL. Client-facing communication skills and experience creating detailed presentations and documentation are essential. A Bachelor's degree in Computer Science or a related field is required.
Requirements
- 8+ years as a hands-on Data Engineer designing and implementing data solutions
- Programming expertise in Java, Python and/or Scala
- Core cloud data platforms including Snowflake, AWS, Azure, Databricks and GCP
- SQL and the ability to write, debug, and optimize SQL queries
- Client-facing written and verbal communication skills and experience
- 4-year Bachelor's degree in Computer Science or a related field
Responsibilities
- Develop end-to-end technical solutions into production, ensuring performance, security, scalability, and robust data integration
- Lead and/or mentor other engineers
- Create and deliver detailed presentations
- Develop detailed solution documentation (e.g., including POCs and roadmaps, sequence diagrams, class hierarchies, logical system views, etc.)
Preferred Qualifications
- Production experience in core data platforms: Snowflake, AWS, Azure, GCP, Hadoop, Databricks
- Cloud and Distributed Data Storage: S3, ADLS, HDFS, GCS, Kudu, ElasticSearch/Solr, Cassandra or other NoSQL storage systems
- Data integration technologies: Spark, Kafka, event/streaming, Streamsets, Matillion, Fivetran, NiFi, AWS Data Migration Services, Azure DataFactory, Informatica Intelligent Cloud Services (IICS), Google DataProc or other data integration technologies
- Multiple data sources (e.g. queues, relational databases, files, search, API)
- Complete software development lifecycle experience including design, documentation, implementation, testing, and deployment
- Automated data transformation and data curation: dbt , Spark, Spark streaming, automated pipelines
- Workflow Management and Orchestration : Airflow, AWS Managed Airflow, Luigi, NiFi
Benefits
- Remote-First Workplace
- Medical Insurance for Self & Family
- Medical Insurance for Parents
- Term Life & Personal Accident
- Wellness Allowance
- Broadband Reimbursement
- Continuous learning and growth opportunities to enhance your skills and expertise
- Other benefits include paid certifications, professional development allowance, and bonuses for creating for company-approved content