Data Engineer, BI
Allata
πRemote - United States
Please let Allata know you found this job on JobsCollider. Thanks! π
Job highlights
Summary
Join Allata, a rapidly growing technology consulting firm, as a skilled Data Engineer. Contribute to transformative enterprise data platform projects, focusing on developing data pipelines and logic engines for data management and product modeling. You will design, build, deploy, and optimize data products for large enterprise clients across various industries. This role demands expertise in data architecture, warehousing, ETL/ELT processes, and proficiency in SQL, Stored Procedures, and PySpark. You will collaborate with cross-functional teams and leverage modern data tools like Databricks and Azure Data Fabric. The ideal candidate is a problem-solver with strong communication and teamwork skills.
Requirements
- Current knowledge of and using modern data tools like (Databricks,FiveTran, Data Fabric and others); Core experience with data architecture, data integrations, data warehousing, and ETL/ELT processes
- Applied experience with developing and deploying custom whl and or in session notebook scripts for custom execution across parallel executor and worker nodes
- Applied experience in SQL, Stored Procedures, and Pyspark based on area of data platform specialization
- Strong knowledge of cloud and hybrid relational database systems, such as MS SQL Server, PostgresSQL, Oracle, Azure SQL, AWS RDS, Aurora or a comparable engine
- Strong experience with batch and streaming data processing techniques and file compactization strategies
Responsibilities
- Collaborate in defining the overall architecture of the solution. This includes knowledge of modern Enterprise Data Warehouse and Data Lakehouse architectures that implement Medallion or Lamda architectures
- Design, develop, test, and deploy processing modules to implement data-driven rules using SQL, Stored Procedures, and Pyspark
- Understand and own data product engineering deliverables relative to a CI-CD pipeline and standard devops practices and principles
- Build and optimize data pipelines on platforms like Databricks, SQL Server, or Azure Data Fabric
Preferred Qualifications
- Automation experience with CICD pipelines to support deployment and integration workflows including trunk-based development using automation services such as Azure DevOps, Jenkins, Octopus
- Advanced proficiency in Pyspark for advanced data processing tasks
- Advance proficiency in spark workflow optimization and orchestration using tools such as Asset Bundles or DAG (Directed Acyclic Graph) orchestration
- Ability to identify, troubleshoot, and resolve complex data issues effectively
- Strong teamwork, communication skills and intellectual curiosity to work collaboratively and effectively with cross-functional teams
- Commitment to delivering high-quality, accurate, and reliable data products solutions
- Willingness to embrace new tools, technologies, and methodologies
- Innovative thinker with a proactive approach to overcoming challenges
Share this job:
Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.
Similar Remote Jobs
- π°$52kπSlovakia
- πCzech Republic
- πPoland
- π°$195k-$247kπUnited States
- πUnited States
- πIndia
- π°$142k-$282kπUnited States, Canada
- πWorldwide
- π°$100k-$180kπUnited States
Please let Allata know you found this job on JobsCollider. Thanks! π