Description

DATA/SOFTWARE ENGINEER

RESPONSIBILITIES

Ø Develop Data Pipelines using PySpark libraries.

Ø Design and implement unit/component/integration tests

Ø Collaborate with other engineers on creating common code bas using Github SCM

Ø Take care of the non-functional requirement like Security, Compliance, performance and maintenance.

Ø Learn new technologies/tools and be able to support that in quick span of time

REQUIRED

Ø Bachelor's degree in statistics, business, economics, mathematics, computer science, engineering, research, or a related field.

Ø 7+ years of industry experience, 3+ years of relevant big data experience

Ø Demonstrated experience and ability to deliver results on multiple projects in a fast-paced, agile environment

Ø Strong programming experience in Python, good knowledge of PySpark

Ø Experience working with Hadoop and Big Data processing frameworks (Spark, Hive, Nifi, Spark-Streaming, Flink, etc.)

 

Ø Knowledge of SCM Tools – Github, Bitbucket, etc

Ø Experience working with NoSQL data stores such as HBase, DynamoDB, etc.

Ø Experience with SQL and SQL Analytical functions

Ø Experience building domain-driven Microservices

Ø Experience working with structured and unstructured datasets, including CSV, TXT, JSON, and XML

Ø Experience provisioning RESTful API’s to enable real-time data consumption

 

 

Ø Ability to collaborate with key partners to understand, identify, and capture requirements

Ø Ability to work in cross-functional teams following Agile Methodologies (SAFe, SCRUM)

Ø A flexible, dynamic personality who works well partnering/collaborating in a matrixed team environment

Ø Ability to work independently requiring minimal direction, execute large projects following established guidelines/procedures and take the initiative to explore and solve problems

Ø Desire to work collaboratively with your teammates to come up with the best solution to a problem

Ø Strong desire to learn and share knowledge with others

PREFERRED

Ø Master’s degree (M.A., M.S. or MBA) in analytics, information technology, computer science, economics, statistics, engineering, research or a related field

Ø 5+ years of relevant big data experience

Ø Proven skill in using Agile/LEAN Methodologies

Ø Strong SQL skills, experience working with large datasets, exceptional analytical and problem-solving skills, data modeling recommendations

Ø Experience provisioning RESTful API’s to enable real-time data consumption

Ø Experience working with Big Data streaming services such as Kinesis, Kafka, etc.

Ø Experience with Agile and LEAN methodologies a plus