Senior or Staff Software Engineer, Data Infrastructure

TRM Labs
Summary
Join TRM Labs, a blockchain intelligence company fighting crime and building a safer world, as a Data Infrastructure Engineer. You will design and maintain petabyte-scale databases, build and optimize data pipelines, diagnose and tune SQL queries, and own key infrastructure initiatives. Collaborate with data scientists, engineers, and product managers to deliver reliable data workflows. This role requires 5+ years of experience in data engineering and expertise with at least one RDBMS (Postgres, MySQL, or SQL Server) at production scale. You'll need strong SQL skills, experience designing data models, and familiarity with data pipeline frameworks. The team is remote-first, async-friendly, and values speed and efficiency. TRM offers competitive compensation, equity, and generous benefits including PTO, holidays, and parental leave.
Requirements
- 5+ years of experience in data engineering, analytics infrastructure, or backend systems with RDBMS depth
- Experience implementing and maintaining database security measures, including access control, encryption, and compliance with security frameworks and standards
- Proven expertise with at least one of: Postgres, MySQL, or SQL Server at production scale (e.g., TB-scale datasets, concurrency, replication, tuning)
- Strong command of SQL reasoning—you know how to debug, explain, and optimize queries, not just write them
- Experience designing and evolving data models (normalized and denormalized) to support analytical or operational use cases
- Familiarity with data pipeline frameworks (e.g., Airflow, DBT, custom orchestration)
- Systems thinking and ownership mindset—you’re comfortable solving ambiguous, cross-functional problems from end to end
Responsibilities
- Design and maintain petabyte scale high-performance databases and data models that support real-time investigations and analytics use cases
- Build and optimize production data pipelines—batch and streaming—that transform large-scale blockchain datasets
- Diagnose and tune complex SQL queries under heavy load, working closely with product and research teams
- Own key infrastructure initiatives—from query optimization and index strategy to storage optimization and system resilience
- Collaborate cross-functionally to deliver reliable and impactful data workflows end-to-end
- Build scalable engines to optimize routine scaling and maintenance tasks like create self-serve automation for creating new pgbouncer, scaling disks, scaling/updating of clusters, etc
- Enable tasks to be faster next time and reducing dependency on a single person
- Identify ways to compress timelines using 80/20 principle . For instance, what does it take to be operational in a new environment? Identify the must have and nice to haves that are need to deploy our stack to be fully operation. Focus on must haves first to get us operational and then use future milestones to harden for customer readiness. We think in terms of weeks and not months
- Identify first version, a.k.a., " skateboards " for projects. For instance, build an observability dashboard within a week. Gather feedback from stakeholders after to identify more needs or bells and whistles to add to the dashboard
Benefits
- Paid Time Off
- Holidays
- Parental Leave
- Remote work