Job Description
Responsibilities: -
· Design and implement the hashtag#datamodeling, hashtag#dataingestion and hashtag#dataprocessing for various hashtag#datasets
· hashtag#Design, hashtag#develop and maintain hashtag#ETL hashtag#Framework for various new data source
· Ability to migrate the existing hashtag#Talend hashtag#ETL workflow into new hashtag#ETL framework using hashtag#AWS hashtag#Glue/ hashtag#EMR, hashtag#PySpark and/or hashtag#datapipeline using hashtag#python.
· Develop and execute hashtag#adhoc hashtag#dataingestion to support business analytics.
· Proactively interact with vendors for any questions and report the status accordingly
· Explore and evaluate the tools/service to support business requirement
· Ability to learn to create a data-driven culture and impactful hashtag#datastrategies.
· Aptitude towards learning new technologies and solving complex problem.
Required Skills:
· Minimum of bachelor’s degree. Preferably in Computer Science, Information system, Information technology.
· Minimum 5 years of experience on hashtag#cloud platforms such as hashtag#AWS, hashtag#Azure, hashtag#GCP.
· Minimum 5 year of experience in hashtag#AmazonWebServices like hashtag#VPC, hashtag#S3, hashtag#EC2, hashtag#Redshift, hashtag#RDS, hashtag#EMR, hashtag#Athena, hashtag#IAM, hashtag#Glue, hashtag#DMS, hashtag#Datapipeline & hashtag#API, etc.
· Minimum of 5 years of experience in hashtag#ETL and hashtag#dataengineering using hashtag#Python, hashtag#AWS hashtag#Glue, hashtag#AWS hashtag#EMR / hashtag#PySpark and hashtag#Talend.
· Minimum 5 years of experience in hashtag#SQL, hashtag#Python, and source control.
· Experience in hashtag#PostgreSQL, hashtag#SQLServer, hashtag#MySQL & hashtag#Oracle databases.
· Experience in hashtag#MPP such as hashtag#AWSRedshift and hashtag#EMR.
· Experience in distributed programming with hashtag#Python, hashtag#UnixScripting, hashtag#MPP, hashtag#RDBMS databases for hashtag#dataintegration.
· Experience building distributed high-performance systems using hashtag#Spark/ hashtag#PySpark, hashtag#AWS hashtag#Glue and developing applications for loading/streaming data into databases, hashtag#Redshift
· Experience in hashtag#Agile methodology.
· Proven skills to write technical specifications for data extraction and good quality code.
· Experience in analytic tools hashtag#Tableau/ hashtag#Qlik.
· Experience with big data processing techniques using hashtag#Sqoop, hashtag#Spark, hashtag#hive is additional
Bachelor's degree in Computer Science