What You'll Do
Design and implement data monitoring pipelines to proactively identify and resolve data quality issues, potentially impacting downstream products
Collaborate with stakeholders to define requirements, develop metrics for data pipeline quality, negotiate data quality SLAs on behalf of downstream data product owners, and create monitoring solutions using Python, Spark, and Airflow
Innovate and develop new methodologies to enhance access to trustworthy data, accelerating the value provided by the product data team
What You Bring
Bachelor’s Degree in Computer Science, Engineering, or a related STEM field, with a focus on data processing
A Master’s Degree is equivalent to 2 years of experience
A Ph.D. counts as 5 years of experience
At least 8 years of experience in Data Engineering or a similar role, with a proven track record of working with big data pipelines and analytics
Minimum of 5 years of hands-on experience with SQL in scalable data warehouses (e.g., Bigquery, Snowflake)
Proficiency in cloud technologies, preferably GCP and/or AWS
Expertise with Apache AirfloW
Over 8 years of coding experience in Python, Java, or equivalent programming language, beyond an undergraduate degree
Deep understanding of Distributed Systems and Effective Data Management
Expertise with CI/CD tools (e.g., Jenkins
ANY GRADUATE