Skillsets we are looking for:
8 to 11 years of working experience in Data Engineering.
6+ years experience in PySprak.
5+ years experience in AWS Glue.
5+ years experience in AWS Redshift.
5+ years exp in AWS CI/CD pipeline like code build, code commit, code deploy and code pipeline.
Experience in ETL tools.
Strong proficiency in AWS services such as S3, EC2, EMR, SNS, Lambda, Step Functions and Event bridge.
Experience implementing automated testing platforms like PyTest.
Strong proficiency in Python, Hadoop, Spark and or PySpark is required.
Skill of writing clean, readable, commented and easily maintainable code.
Understanding of fundamental design principles for building a scalable solution.
Skill for writing reusable libraries.
Proficiency in understanding code versioning tools such as Git, SVN, TFS etc.,
Bachelor's degree or higher.
Bachelor's degree