Description

 Interact with business stake holders and designers to implement to understand business

requirements.

 Ability to perform impact assessments on a vast data store to ensure existing data pipeline isn’t wrecked and uncover insights.

 Translate complex functional and technical requirements into detailed design.

 Project Development and Implementation experience working in a Hadoop Distributed File Hadoop System

 Designing, building, installing, configuring, and supporting in a Hadoop based environment.

 Ingestion of Complex data sets into Hadoop environment through various techniques

(Ingestion via Spark / Hive / Sqoop techniques)

 Transform data using Spark with Scala

 Managing and deploying Hive objects.

 Should have performed Unit/System Testing to ensure code quality.

 Good to have Teradata Knowledge

 Must have working experience in IntelliJ IDEA, Autosys Job scheduling and should work

seamlessly in WinSCP, Putty and Unix

 Must have working knowledge in utilizing GitHub, CI-CD pipelines like TeamCity or Jenkins in

productionizing the code.

 Maintain Security and data privacy.

 Create scalable and high-performance web services for data tracking.

 High-speed querying capability

 Test prototypes and oversee handover to operational teams.

Education

Any Graduate