Description

Mandatory Skills:
ETL Process with PySpark, Spark SQL, SQL tuning, data validation, dimensional and relational data modeling
Building data architectures and data pipelines in support of analytics
Big-data technologies such as Hadoop, SparkML, etc.
Data-related operations (e.g. SQL, UNIX)
Experienced with Agile/SCRUM framework

Responsibilities:
Streamlining and automating current processes(based on xls or Alteryx) for forecasting, cap planning & analytics consumption with PySpark
Convert existing Alteryx workflow into analytics supported formats, define the data and reporting operations processes & performance monitoring metrics. 
Implement rapid prototyping & deployment of existing capability to the target state platform.

Education

Any Graduate