Job Code : EWC - 8
Responsibilities:
·Develop and automate large scale, high-performance data processing systems (batch and/or streaming) to drive Airbnb business growth and improve the product experience.
·Build scalable Spark data pipelines leveraging Airflow scheduler/executor framework
Minimum Requirements:
·8+ years of relevant industry experience
·Demonstrated ability to analyze large data sets to identify gaps and inconsistencies, provide data insights, and advance effective product solutions
·Working knowledge of relational databases and query authoring (SQL).
·Good communication skills, both written and verbal
·Strong experience using ETL framework (ex: Airflow, Flume, Oozie etc.) to build and deploy production-quality ETL pipelines.
·Experience building batch data pipelines in Spark Scala.
·Strong understanding of distributed storage and compute (S3, Hive, Spark)
·General software engineering skills (Java or Python, Github)
ANY GRADUATE