Description

About the job

AWS cloud in S3/ATHENA/GLUE/EMR/VPC.

Creating and Managing Data Pipelines and Streamsets in KAFKA

5+ yrs of Proven experience as a PySpark Developer or a related role

Strong programming skills in Scala, Spark or Python

Experience in working with AWS cloud and AWS DocuDB

Familiarity with big data processing tools and techniques

Experience with streaming data platforms

Excellent analytical and problem-solving skills

Strong knowledge in Kafka topics and K-SQL

Extensive Unix/FTP/ File Handling

Strong Hands-on in No SQL databases like Mongo DB

Experience with Agile Methodology/JIRA/Confluence

Hands-on in Handling all Processors.

Responsibilities

Extensive knowledge in understanding Data and creating Data Pipelines.

Interact with Business to understand the requirement on a daily evolving need and adapt to the needs.

Handle streaming data platforms .

Able to handle all the database storage and retrieval mechanisms for the Mongo DB on and extensive experience in handling.

Able to understand the business requirements, tech stack involved and guide the team towards Integral solution approach.

Provide extensive development support to the Team in achieving the Project Milestones.

Dedicated and committed approach with excellent communication skills.

Education

Any Graduate