Full Stack Engineer, Data Products

Stripe Logo

Stripe

πŸ“Remote - United States

Summary

Join Stripe's Data Movement team as a Staff Software Engineer and build and operate a large-scale read-only ecosystem extending Stripe's primary online database. You will design, build, and maintain innovative data solutions, define architectural strategies, and ensure operational excellence of the data platform. Collaborate with various teams to support their initiatives and contribute to open-source software improvement. This is a high-visibility role requiring strong technical leadership and a team-first mentality. You'll work with technologies like Flink, Spark, Iceberg, Trino, Airflow, and Kafka at a massive scale. The role demands expertise in large-scale data pipelines and distributed systems.

Requirements

  • Has very strong operational experience with large-scale, high-availability data pipelines from design, to execution, and safe change management
  • Expertise in Flink, Spark (especially structured streaming), Airflow, Python, Java, SQL, Trino, and API design is a plus
  • Has experience developing, maintaining, and debugging distributed systems built with open source tools
  • Has experience building infrastructure-as-a-product with a strong focus on users needs
  • Has strong collaboration and communication skills, and can comfortably interact with both technical and non-technical participants
  • Has the curiosity to continuously learn about new technologies and business processes
  • Is energized by delivering effective, user-first solutions through creative problem-solving and collaboration

Responsibilities

  • Design, build, and maintain innovative next-generation or first-generation versions of solutions and products, with an emphasis on reliability, scalability, and efficiency
  • Define and design opinionated architectural strategies, policies, and abstractions that build a long term sustainable solution for internal Stripes, that will in turn enhance the experience of millions of Stripe users
  • Ensure operational excellence and enable a highly available & reliable Data Movement platform across streaming and batch workloads, delivering a world-class data and query ecosystem
  • Collaborate nimbly with high-visibility teams and their stakeholders to support their key initiatives - navigating various types of organizational complexity with the support of your own leadership team
  • Connect your work with improvements in the usability and reliability of Open Source Software (OSS) like Apache Airflow, Iceberg, Spark and contribute back to the OSS community

Preferred Qualifications

  • Has experience writing production-level code in expertise in Scala, Spark, Flink, Iceberg, Airflow, Python, Java, and SQL is a plus
  • Experience packaging and deploying code into cloud-based environments (AWS, GCP, Azure) with tools including Bazel, Docker Containers, etc
  • Has experience designing APIs or building developer platforms
  • Has experience optimizing the end to end performance of distributed systems
  • Has experience with scaling distributed systems in a rapidly moving environment
  • Has experience working with data pipelines

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.