Responsibilities:
- Design, implement and support an analytical data infrastructure
- Managing AWS resources including EC2, EMR, S3, Glue, Redshift, etc.
- Interface with other technology teams to extract, transform, and load data from a wide variety of data sources using SQL and AWS big data technologies
- Explore and learn the latest AWS technologies to provide new capabilities and increase efficiency
- Collaborate with DataScientists and Business IntelligenceEngineers (BIEs) to recognize and help adopt best practices in reporting and analysis
- Maintain internal reporting platforms/tools including troubleshooting and development. Interact with internal users to establish and clarify requirements in order to develop report specifications.
- Write advanced SQL queries and Python code to develop solutions
Required Skills:
Education Requirement: Bachelor's or Master's degrees in Computer Science, Information Science, Electrical Engineering, or their foreign equivalents in education.
Experience:
- Work experience with ETL, Data Modelling, and Data Architecture.
- Expert-level skills in writing and optimizing SQL.
- Experience with big data technologies such as Hadoop/Hive/Spark.
- Solid Linux skills.
- Experience operating very large data warehouses or data lakes.
- Expertise in ETL optimization, designing, coding, and tuning big data processes using Apache Spark or similar technologies.
- Experience with building data pipelines and applications to stream and process datasets at low latencies.