Description

Job Description
Responsibilities: -
·        Design and implement the hashtag#datamodeling, hashtag#dataingestion and hashtag#dataprocessing for various hashtag#datasets
· hashtag#Design, hashtag#develop and maintain hashtag#ETL hashtag#Framework for various new data source
·        Ability to migrate the existing hashtag#Talend hashtag#ETL workflow into new hashtag#ETL framework using hashtag#AWS hashtag#Glue/ hashtag#EMR, hashtag#PySpark and/or hashtag#datapipeline using hashtag#python.
·        Develop and execute hashtag#adhoc hashtag#dataingestion to support business analytics.
·        Proactively interact with vendors for any questions and report the status accordingly
·        Explore and evaluate the tools/service to support business requirement
·        Ability to learn to create a data-driven culture and impactful hashtag#datastrategies.
·        Aptitude towards learning new technologies and solving complex problem.
Required Skills:
·        Minimum of bachelor’s degree. Preferably in Computer Science, Information system, Information technology.
·        Minimum 5 years of experience on hashtag#cloud platforms such as hashtag#AWS, hashtag#Azure, hashtag#GCP.
·        Minimum 5 year of experience in hashtag#AmazonWebServices like hashtag#VPC, hashtag#S3, hashtag#EC2, hashtag#Redshift, hashtag#RDS, hashtag#EMR, hashtag#Athena, hashtag#IAM, hashtag#Glue, hashtag#DMS, hashtag#Datapipeline & hashtag#API, etc.
·        Minimum of 5 years of experience in hashtag#ETL and hashtag#dataengineering using hashtag#Python, hashtag#AWS hashtag#Glue, hashtag#AWS hashtag#EMR / hashtag#PySpark and hashtag#Talend.
·        Minimum 5 years of experience in hashtag#SQL, hashtag#Python, and source control.
·        Experience in hashtag#PostgreSQL, hashtag#SQLServer, hashtag#MySQL & hashtag#Oracle databases.
·        Experience in hashtag#MPP such as hashtag#AWSRedshift and hashtag#EMR.
·        Experience in distributed programming with hashtag#Python, hashtag#UnixScripting, hashtag#MPP, hashtag#RDBMS databases for hashtag#dataintegration.
·        Experience building distributed high-performance systems using hashtag#Spark/ hashtag#PySpark, hashtag#AWS hashtag#Glue and developing applications for loading/streaming data into databases, hashtag#Redshift
·        Experience in hashtag#Agile methodology.
·        Proven skills to write technical specifications for data extraction and good quality code.
·        Experience in analytic tools hashtag#Tableau/ hashtag#Qlik.
·        Experience with big data processing techniques using hashtag#Sqoop, hashtag#Spark, hashtag#hive is additional

Education

Bachelor's degree in Computer Science