Data Infrastructure Engineer

Bastion Logo

Bastion

πŸ“Remote - Worldwide

Summary

Join Bastion as a Data Infrastructure Engineer and build and maintain critical data infrastructure, including ingestion, analysis, and reporting pipelines. You will work with cross-functional teams, select appropriate technologies, manage vendor relationships, and foster a data-driven culture. This foundational role requires architecting and building real-time and batch data pipelines, establishing strong security and privacy controls, and working with various data sources. You will also be responsible for data governance, documentation, and identifying platform needs. The ideal candidate possesses extensive experience in data analytics, pipeline technologies, and cloud-based data lakes. Bastion values ownership, a customer-centric mindset, attention to detail, teamwork, and continuous improvement.

Requirements

  • 5+ years of professional engineering and data analytics experience, startup experience a plus
  • Strong proficiency and comfort using SQL and Python to perform complex data analysis
  • Recent experience building automation tooling and pipelines using a general purpose language such as Python, Golang, and/or Typescript
  • Experience with modern data pipeline and warehouse technologies (e.g. Snowflake, Databricks, Apache Spark, AWS Glue)
  • Strong proficient writing declarative data models and transformations using modern technologies (e.g. dbt)
  • Experience building and maintaining cloud-based data lakes
  • Prior experience with integrating real-time data streaming technologies (e.g. Kafka, Spark)
  • Prior experience with configuring and maintaining modern data orchestration platforms (e.g. Airflow)
  • Comfort with infrastructure-as-code tooling (e.g. Terraform) and container orchestration platforms (e.g. Kubernetes)
  • Strong preference to keep things simple, ship fast, and avoid overengineering
  • Self-driven and ability to work autonomously

Responsibilities

  • Architect, build, and maintain modern and robust real-time and batch data analytics pipelines
  • Develop and maintain declarative data models and transformations
  • Implement data ingestion integrations for streaming and traditional sources such as Postgres, Kafka, and DynamoDB
  • Deploy and configure BI tooling for data analysis
  • Work closely with product, finance, legal, and compliance teams to build dashboards and reports to support business operations, regulatory obligations, and customer needs
  • Establish, communicate, and enforce data governance policies
  • Document and share best practices with regards to schema management, data integrity, availability, and security
  • Protect and limit access to sensitive data by implementing a secure permissioning model and establishing data masking and tokenization processes
  • Identify and communicate data platform needs, including additional tooling and staffing
  • Work with cross-functional teams to define requirements, plan projects, and execute on the plan

Preferred Qualifications

Professional Web3 / Crypto experience is a plus

Share this job:

Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.

Similar Remote Jobs