Senior Or Staff Software Engineer, Data Platform

TRM Labs Logo

TRM Labs

πŸ’΅ $200k-$230k
πŸ“Remote - United States

Summary

Join TRM Labs, a blockchain intelligence company fighting crime globally. As a Senior Data Engineer, you will design, implement, and scale core components of our lakehouse architecture on GCP. You'll work with cutting-edge tools like Apache Spark, Trino, and Snowflake, focusing on data modeling, ingestion, and query optimization. This role requires 5+ years of experience in data or software engineering with a focus on distributed data systems and cloud-native architectures. TRM offers a remote-first, collaborative environment with competitive compensation, including a base salary range of $200,000-$230,000 and equity participation. The company prioritizes a growth mindset and provides development opportunities.

Requirements

  • 5+ years of experience in data or software engineering, with a focus on distributed data systems and cloud-native architectures
  • Proven experience building and scaling data platforms on GCP, including storage, compute, orchestration, and monitoring
  • Strong command of one or more query engines such as Trino, Presto, Spark, or Snowflake
  • Experience with modern table formats like Apache Hudi, Iceberg, or Delta Lake
  • Exceptional programming skills in Python, as well as adeptness in SQL or SparkSQL
  • Hands-on experience orchestrating workflows with Airflow and building streaming/batch pipelines using GCP-native services

Responsibilities

  • Architect and scale a high-performance data lakehouse on GCP, leveraging services like BigQuery, Dataproc, and Pub/Sub
  • Design, build, and optimize distributed query engines such as Trino, Spark, or Snowflake to support complex analytical workloads
  • Implement metadata management and data discovery frameworks for governance and observability using technologies like Apache Hudi, Iceberg, or Delta Lake
  • Develop and orchestrate robust ETL/ELT pipelines using Apache Airflow, Spark, and GCP-native tools (e.g., Dataflow, Composer)
  • Collaborate across departments, partnering with data scientists, backend engineers, and product managers to design and implement
  • Build scalable engines to optimize routine scaling and maintenance tasks like create self-serve automation for creating new pgbouncer, scaling disks, scaling/updating of clusters, etc
  • Enable tasks to be faster next time and reducing dependency on a single person
  • Identify ways to compress timelines using 80/20 principle . For instance, what does it take to be operational in a new environment? Identify the must have and nice to haves that are need to deploy our stack to be fully operation. Focus on must haves first to get us operational and then use future milestones to harden for customer readiness. We think in terms of weeks and not months
  • Identify first version, a.k.a., " skateboards " for projects. For instance, build an observability dashboard within a week. Gather feedback from stakeholders after to identify more needs or bells and whistles to add to the dashboard

Benefits

  • The estimated base salary range for this role is $200,000 - $230,000
  • Additionally, this role may be eligible to participate in TRM’s equity plan
  • Remote-first
  • Work alongside top experts and learn every day
  • Embrace a growth mindset with development opportunities tailored to your role
  • Take on high-impact challenges in a fast-paced, collaborative environment

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.

Similar Remote Jobs