Description

Duties/Responsibilities:
⚫ Design, build, and maintain data pipelines and ETL processes to ensure reliable and timely data ingestion, transformation, and delivery
• Develop and optimize data models, schemas, and databases to support web applications, data analysis, and reporting needs.
Designs logical data models and implementation of physical database structures
Implement data quality and validation processes to ensure the accuracy, completeness, and consistency of data Implement data security and privacy measures to safeguard sensitive information.
Collaborate with developers and analysts to provide them with the necessary data sets and ensure smooth data access for their analysis and analytical needs.
⚫ Document data engineering processes, best practices, and technical specifications
• Monitor and troubleshoot data pipelines, identifying and resolving data processing, performance, and reliability issues.
• Coordinate with internal teams to understand business requirements
Collaborate with quality assurance teams to identify and resolve defects
Follow automated continuous delivery and deployment practices
Education:
⚫ Bachelor's Degree from an accredited college or university in the specific discipline required by the State.
General Experience:
Minimum 5+ years ETL coding experience
ETL Developer
INNOSOFT
Proficiency in programming languages such as Python and SQL for data processing and automation Experience with distributed computing frameworks like Apache Spark or similar technologies Experience with AWS data environment, primarily EMR, Glue, S3, DocumentDB, Redshift, RDS, Athena, etc. Experience with data warehouses/RDBMS like Redshift and NoSQL data stores such as DocumentDB, DynamoDB, OpenSearch, etc
• Experience in building data lakes using AWS Lake Formation
• Experience with workflow orchestration and scheduling tools like AWS Step Functions, AWS MWAA, etc.. Strong understanding of relational databases (including tables, views, indexes, table spaces)
• Experience with source control tools such as GitHub and related CI/CD processes
• Ability to analyze a company's data needs
Strong problem-solving skills
Experience with the SDLC and Agile methodologies

Certifications:
♦️ Certified Data Management Professional (CDMP)
♦️ IBM Certified Solution Developer 
♦️ SAS Certified Data Integration Developer for SAS 9.
♦️ Informatica Certified Professional (ICP).
 

Education

Bachelor's degree