Senior Data Engineer

phData Logo

phData

πŸ“Remote - India

Summary

Join phData, a leading innovator in the modern data stack, partnering with major cloud data platforms. We are a remote-first global company committed to fostering a community of technological curiosity and ownership. We hire top performers and offer autonomy to deliver results. phData is an award-winning workplace seeking top talent to expand our network of skilled professionals. We encourage applications even if no role is currently available; your information will be added to our system for future opportunities. We celebrate diversity and are committed to creating an inclusive environment for all employees.

Requirements

  • Demonstrate programming expertise in Java, Python, and/or Scala
  • Show proficiency in core cloud data platforms, including Snowflake, AWS, Azure, Databricks, and GCP
  • Write, debug, and optimize SQL queries
  • Possess client-facing written and verbal communication skills and experience
  • Create and deliver detailed presentations
  • Produce detailed solution documentation (e.g., POCs and roadmaps, sequence diagrams, class hierarchies, logical system views, etc.)
  • Hold a 4-year Bachelor's degree in Computer Science or a related field
  • Have at least 4+ years of experience as a Software Engineer, Data Engineer, or Data Analyst

Responsibilities

Develop end-to-end technical solutions into production, ensuring performance, security, scalability, and robust data integration

Preferred Qualifications

  • Possess production experience in core data platforms: Snowflake, AWS, Azure, GCP, Hadoop, Databricks
  • Show expertise in Cloud and Distributed Data Storage: S3, ADLS, HDFS, GCS, Kudu, ElasticSearch/Solr, Cassandra, or other NoSQL storage systems
  • Have experience with Data integration technologies: Spark, Kafka, event/streaming, Streamsets, Matillion, Fivetran, NiFi, AWS Data Migration Services, Azure DataFactory, Informatica Intelligent Cloud Services (IICS), Google DataProc, or other data integration technologies
  • Demonstrate experience with Multiple data sources (e.g., queues, relational databases, files, search, API)
  • Have complete software development lifecycle experience, including design, documentation, implementation, testing, and deployment
  • Show expertise in Automated data transformation and data curation: dbt, Spark, Spark streaming, automated pipelines
  • Have experience with Workflow Management and Orchestration: Airflow, AWS Managed Airflow, Luigi, NiFi

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.

Similar Remote Jobs