Remote SME Data Engineer
Shuvel
πRemote - United States
Please let Shuvel know you found this job on JobsCollider. Thanks! π
Job highlights
Summary
The job is for a SME Data Engineer at CBP, responsible for designing, developing, and maintaining scalable data pipelines and architectures. The position requires extensive experience with SQL, data warehousing concepts, distributed computing principles, and cloud environments. Preferred qualifications include 5+ years of experience in developing complex Oracle PL/SQL packages and certifications in relevant technologies.
Requirements
- Experience with relational databases and knowledge of query tools and/or BI tools like Power BI or OBIEE and data analysis tools
- Extensive experience with SQL and proficiency in writing complex queries
- Solid understanding of data warehousing concepts and platforms such as Oracle and cloud-based solutions
- Strong experience in automating ETL jobs via UNIX/LINUX shell scripts and CRON jobs
- Demonstrate a strong practical understanding of data warehousing from a production relational database environment
- Strong experience using analytic functions within Oracle or similar tools within non-relational (MongoDB, Cassandra etc.) database systems
- Strong understanding of distributed computing principles and experience with frameworks like Apache Spark
- Hands-on-experience with data lake architectures and technologies in a cloud environment
- Experience with Atlassian suite of tools such as Jira and Confluence
- Knowledge of Continuous Integration & Continuous Development tools (CI/CD)
- Must be able to multitask efficiently and progressively and work comfortably in an ever-changing data environment
- Must work well in a team environment as well as independently
- Excellent verbal/written communication and problem-solving skills; ability to communicate information to a variety of groups at different technical skill levels
Responsibilities
- Design, develop, and maintain scalable data pipelines and architectures to support ETL/ELT processes
- Utilize strong SQL skills to perform complex data transformations and optimize database queries
- Build comprehensive datasets by aggregating data sourced from various relational databases
- Collaborate with cross-functional teams to understand business requirements and translate them into technical solutions
- Assist with the implementation of data migration/pipelines from on-prem to cloud/non-relational storage platforms
- Leverage distributed computing frameworks like Apache Spark to process large volumes of data efficiently
- Utilize data analysis, problem-solving, investigative, and creative thinking skills to handle extremely large datasets
- Respond to data queries/analysis requests from various groups within an organization
- Troubleshoot data-related issues, identify root causes, and implement solutions to ensure data integrity and accuracy
- Implement best practices for data governance, security, and quality supporting the core business applications
- Responsible for data engineering source code control using GitLab
Preferred Qualifications
- 5+ years of experience in developing, maintaining, and optimizing complex Oracle PL/SQL packages to aggregate transactional data for consumption by data science/machine learning applications
- 10+ years of experience in working in data engineering, with a focus on building and optimizing data pipelines and architectures. Must have full life cycle experience in design, development, deployment, and monitoring
- Experience with one or more relational database systems such as Oracle, MySQL, Postgres, SQL server, with heavy emphasis on Oracle
- Extensive experience with cloud platforms (e.g. AWS, Google Cloud, etc) and cloud based ETL/ELT tools
- Experience with Amazon services such as S3, Redshift, EMR and Scala
- Experience with migrating on-prem legacy database objects and data to the Amazon S3 cloud environment
- Experience or familiarity with data science/machine learning and development experience for supervised and unsupervised learning with structure and unstructured datasets
- Certifications in relevant technologies (e.g. AWS Certified Big Data, Google Professional Data Engineer) are a plus
Share this job:
Disclaimer: Please check that the job is real before you apply. Applying might take you to another website that we don't own. Please be aware that any actions taken during the application process are solely your responsibility, and we bear no responsibility for any outcomes.
Similar Remote Jobs
- πCanada
- πRomania
- πBulgaria
- πUnited States
- πWorldwide
- πWorldwide
- π°$172k-$215kπUnited States
- πIndia
- π°$124k-$220kπUnited States
Please let Shuvel know you found this job on JobsCollider. Thanks! π