Description

About the job
General Duties/Responsibilities:

(May include but are not limited to)

•Interfacing with business customers, gathering requirements and developing new datasets in data platform

•Building and migrating the complex ETL pipelines fromon-premisee system to cloud and Hadoop/Spark to make the system grow elastically

•Identifying the data quality issues to address them immediately to provide great user experience

•Extracting and combining data from various heterogeneous data sources

•Designing, implementing and supporting a platform that can provide ad-hoc access to large datasets

•Modelling data and metadata to support machine learning and AI

Minimum Requirements:

To perform this job successfully, an individual must be able to perform each essential duty satisfactorily. The requirements listed below are representative of the knowledge, skill, and/or ability required. Reasonable accommodations may be made to enable individuals with disabilities to perform the essential functions.

 

Minimum Experience

3+ years relevant experience in cloud based data engineering.

Demonstrated ability in data modeling, ETL development, and data warehousing.

Data Warehousing Experience with SQL Server, Oracle, Redshift, Teradata, etc.

Experience with Big Data Technologies (NoSQL databases, Hadoop, Hive, Hbase, Pig, Spark, Elasticsearch etc.)

Experience in using Scala, Python, .net, Java and/or other data engineering

Knowledge and experience of SQL Sever and SSIS.

Excellent communication, analytical and collaborative problem-solving skills

API development experience is a plus

Education

Any Graduate