Description

Job Description

Required:

AWS, Kafka, Python, Terraform. Stream Processing application development experience using Kafka Streams and/or Flink APIs. Experience with Postgres, MySQL, DocumentDB.

Nice To Have

Flink, EMR, Hudi, AWS-Redshift, ECS/EKS, Glue Catalog – at least Solution Architect Associate, Developer, or Data Engineering certification preferred.

Roles & Responsibilities

Support or collaborate with application developers, database architects, data analysts and data scientists to ensure optimal data delivery architecture throughout ongoing projects/operations. 
Design, build, and manage analytics infrastructure that can be utilized by data analysts, data scientists, and non-technical data consumers, which enables functions of the big data platform for Analytics. 
Develop, construct, test, and maintain architectures, such as databases and large-scale processing systems that help analyse and process data in the way the Analytics organization requires. 
Develop highly scalable data management interfaces, as well as software components by employing programming languages and tools. 
Design, document, build, test and deploy data pipelines that assemble large complex datasets from various sources and integrate them into a unified view. 
Identify, design, and implement operational improvements: automating manual processes, data quality checks, error handling and recovery, re-designing infrastructure as needed. 
Create data models that will allow analytics and business teams to derive insights about customer behaviours. 
Build new data pipelines, identify existing data gaps and provide automated solutions to deliver analytical capabilities and enriched data to applications. 
Responsible for obtaining data from the System of Record and establishing batch or real-time data feed to provide analysis in an automated fashion. 
Develop techniques supporting trending and analytic decision-making processes 
Apply technologies for responsive front-end experience 
Ensure systems meet business requirements and industry practices 
Research opportunities for data acquisition and new uses for existing data 
Develop data set processes for data modelling, mining and production 
Integrate data management technologies and software engineering tools into existing structures 
Employ a variety of languages and tools (e.g. scripting languages)

What are the top 3 Must Haves for this role?

AWS 
Kafka 
Python

Additional Information

Provides technical direction, guides the team on key technical aspects and responsible for product tech delivery 
Lead the Design, Build, Test and Deployment of component 
Where applicable in collaboration with Lead Developers (Data Engineer, Software Engineer, Data Scientist, Technical Test Lead) 
Understand requirements / use case to outline technical scope and lead delivery of technical solution 
Confirm required developers and skillsets specific to product 
Provides leadership, direction, peer review and accountability to developers on the product (key responsibility) 
Works closely with the Product Owner to align on delivery goals and timing 
Assists Product Owner with prioritizing and managing team backlog 
Collaborates with Data and Solution architects on key technical decisions 
The architecture and design to deliver the requirements and functionality 
 

Education

Any graduate