Description

Responsibilities:

  • Design, implement and support an analytical data infrastructure
  • Managing AWS resources including EC2, EMR, S3, Glue, Redshift, etc.
  • Interface with other technology teams to extract, transform, and load data from a wide variety of data sources using SQL and AWS big data technologies
  • Explore and learn the latest AWS technologies to provide new capabilities and increase efficiency
  • Collaborate with DataScientists and Business IntelligenceEngineers (BIEs) to recognize and help adopt best practices in reporting and analysis
  • Maintain internal reporting platforms/tools including troubleshooting and development. Interact with internal users to establish and clarify requirements in order to develop report specifications.
  • Write advanced SQL queries and Python code to develop solutions

Required Skills:

Education Requirement: Bachelor's or Master's degrees in Computer Science, Information Science, Electrical Engineering, or their foreign equivalents in education.

Experience:

  • Work experience with ETL, Data Modelling, and Data Architecture.
  • Expert-level skills in writing and optimizing SQL.
  • Experience with big data technologies such as Hadoop/Hive/Spark.
  • Solid Linux skills.
  • Experience operating very large data warehouses or data lakes.
  • Expertise in ETL optimization, designing, coding, and tuning big data processes using Apache Spark or similar technologies.
  • Experience with building data pipelines and applications to stream and process datasets at low latencies.


 

Education

Bachelor's or Master's degrees