Data Engineer

PadSplit Logo

PadSplit

πŸ’΅ $135k-$175k
πŸ“Remote - United States

Summary

Join PadSplit as a Data Engineer and build and maintain scalable data infrastructure. This critical role focuses on optimizing data pipelines, ensuring data reliability, and enabling data-driven decision-making. You will design, build, and optimize ETL/ELT pipelines, develop data models, and optimize database performance. Collaboration with data analysts and other teams is essential. The ideal candidate has 5+ years of experience in data engineering and expertise in PostgreSQL, AWS, and data warehousing technologies. This fully remote position offers competitive compensation, benefits, and equity.

Requirements

  • 5+ years of experience in data engineering or a similar role, with a proven track record of designing scalable data solutions
  • Expertise in PostgreSQL, including database management, query optimization, and performance tuning
  • Hands-on experience with AWS cloud services such as S3, Lambda, Glue, Redshift, and IAM
  • Proficiency in data warehousing technologies like Snowflake, Redshift, or BigQuery for cloud-based data storage and analysis
  • Strong skills in data transformation, modeling, and building efficient ETL/ELT pipelines
  • Experience with data visualization tools like Mode, Looker, Tableau, or Hex to support analytics and reporting
  • Knowledge of ElasticSearch or Solr for implementing search indexing and query capabilities
  • Proficiency in SQL and Python, with experience in automation, scripting, and workflow orchestration (e.g., Airflow)
  • Understanding of CI/CD pipelines, infrastructure-as-code principles, and cloud-based deployment practices
  • Strong analytical and problem-solving abilities, with a passion for leveraging data-driven insights to inform decisions

Responsibilities

  • Design, build, and optimize scalable ETL/ELT pipelines to facilitate seamless data ingestion and transformation processes
  • Develop and maintain data models to enable self-service analytics and reporting across the organization
  • Optimize database performance in PostgreSQL, ensuring efficient data storage, retrieval, and query execution
  • Implement and enhance search capabilities using NoSQL technologies like ElasticSearch or Solr to improve data discovery
  • Collaborate with data analysts to create insightful dashboards that support data-driven decision-making
  • Ensure data quality, governance, and security by adhering to best practices in cloud-based data environments
  • Monitor and troubleshoot issues within data pipelines, focusing on optimizing efficiency and reliability
  • Work closely with software engineers and product teams to integrate data solutions into operational workflows and product development

Preferred Qualifications

  • Experience with streaming data solutions like Kafka or Kinesis
  • Knowledge of machine learning pipelines
  • Familiarity with data privacy regulations such as GDPR or CCPA

Benefits

  • Fully remote position
  • Competitive compensation package including an equity incentive plan
  • National medical, dental, and vision healthcare plans
  • Company provided life insurance policy
  • Optional accidental insurances, FSA, and DCFSA benefits
  • Unlimited paid-time (PTO) policy with eleven (11) company-observed holidays
  • 401(k) plan
  • Twelve (12) weeks of paid time off for both birth and non-birth parents
  • The opportunity to do what you love at a company that is at the forefront of solving the affordable housing crisis
  • $135,000 - $175,000 a year

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.