Description

 • Analyze and understand data sources & APIs

• Design and Develop methods to connect & collect data from different data sources

• Design and Develop methods to filter/cleanse the data

• Design and Develop SQL , Hive queries, APIs to extract data from the store

• Work closely with data Scientists to ensure the source data is aggregated and cleansed

• Work with product managers to understand the business objectives

• Work with cloud and data architects to define robust architecture in cloud setup pipelines and work flows

• Work with DevOps to build automated data pipelines

 

Total Experience Required

• 4 <years <10 of relevant experience

• The candidate should have performed client facing roles and possess excellent communication skills

 

Business Domain knowledge: Finance & banking systems, Fraud, Payments

 

Required Technical Skills

• Big Data-Hadoop, NoSQL, Hive, Apache Spark

• Python

• Java & REST

• GIT and Version Control

 

Desirable Technical Skills

• Familiarity with HTTP and invoking web-APIs

• Exposure to machine learning engineering

• Exposure to NLP and text processing

• Experience with pipelines, job scheduling and workflow management

 

Personal Skills

Experienced in managing work with distributed teams

• Experience working in SCRUM methodology

• Proven sense of high accountability and self-drive to take on and see-through big challenges

• Confident, takes ownership, willingness to get the job done

• Excellent verbal communications and cross group collaboration skills

Education

Bachelor's degree in Computer Science