Description

Description:

As a Data Engineer, you will be responsible for the development of complex data sources and pipelines into our data platform (i.e. Snowflake) along with other data applications (i.e. Azure, Terraform, etc.), automation and innovation.

Primary Responsibilities:
• Create & maintain data pipelines using Azure & Snowflake as primary tools
• Create SQL Stored procs and Functions to perform complex transformations
• Understand data requirements and design optimal pipelines to fulfil the use-cases
• Creating logical & physical data models to ensure data integrity is maintained
• Code management, CI/CD pipeline creation & automation using GitHub & GIT Actions
• Tuning and optimizing data processes
• Design and build best in class processes to clean and standardize data
• Code Deployments to production environment, troubleshoot production data issues
• Modelling of big volume datasets to maximize performance for our Business Intelligence & Data Science Team

Qualifications
Required Qualifications:
• Computer Science bachelor's degree or similar
• Min 1-4 years of industry experience as a Hands-on Data engineer
• Excellent communication skills – Verbal and Written
• Excellent knowledge of SQL
• Excellent knowledge of Azure Services such as – Blobs, Functions, Azure Data Factory, Service Principal, Containers, Key Vault, etc.
• Excellent knowledge of Snowflake - Architecture, Features, Best practices
• Excellent knowledge of Data warehousing & BI Solutions
• Excellent Knowledge of change data capture (CDC), ETL, ELT, SCD etc.
• Hands on experience on the following technologies:
o Developing data pipelines in Azure & Snowflake
o Writing complex SQL queries
o Building ETL/ELT/data pipelines using SCD logic
o Query analysis and optimization
• Analytical and problem-solving experience applied to a Big Data datasets
• Data warehousing principles, architecture and its implementation in large environments
• Experience working in projects with agile/scrum methodologies and high performing team(s)
• Knowledge of different data modelling techniques such as Star Schema, Dimensional models, Data vault is an Advantage
• Experience in code lifecycle management and repositories such as GIT & GitHub
• Exposure to DevOps methodology
• Good understanding of Access control and Data masking

Preferred Qualifications:
• Knowledge and experience on Terraform, CI CD Pipelines and automation is an advantage
• Automation and orchestration using ADF
• Create real-time analytics pipelines using Snowpipe Streaming
• Experience developing optimized data models for Viz tool E.g. Tableau, PowerBI is an Advantage
• Exposure and experience in other programming language such as Python, Spark etc. is an advantage
• Hands-on experience of CI CD Pipelines using GIT & GIT Actions
• Understanding of United States Healthcare data and applicable regulations

Education

Any Graduate