Description

  • Design and develop HDFS-based solutions in Hive
  • Program using Hive QL to generate reports
  • Writing Scoop Jobs to Import/Export data from Hadoop
  • Program using Shell script/Python to create automation scripts
  • Knowledge of MapReduce, Spark. Hive, HBASE.  Experience with Eclipse, XML, JSON
  • Program using SQL and SQL/RDBMS utilities
  • Responsible for setting up Hive structures, helping users troubleshoot issues with Hive/Impala/Spark/Sqoop, and migrating the structures between environments.
  • Monitor Hadoop cluster job performance, end-to-end performance tuning of Hadoop clusters and Hadoop Map/Reduce routines against very large data sets.
  • File system management and monitoring, Hadoop HDFS support and maintenance.
  • Manage and review Linux and Hadoop log files.
  • Team with the infrastructure, database, and business analytics teams to guarantee high data quality and availability and to troubleshoot Hadoop issues.

Education

Any Graduate