Description

Required Skills

• 3-6years experience in Hadoop stack and storage technologies, HDFS, MapReduce, Yarn, HIVE, sqoop, Impala , spark, flume, kafka and oozie

• Extensive Knowledge on Bigdata Enterprise architecture (Cloudera preferred)

• Excellent analytical capabilities - Strong interest in algorithms

• Experienced in HBase, RDBMS, SQL, ETL and data analysis

• Experience in No SQL Technologies (ex., Cassandra/ MongoDB, etc )

• Experienced in scripting(Unix/Linux) and scheduling (Autosys)

• Experience with team delivery/release processes and cadence pertaining to code deployment and release

• Research oriented, motivated, pro-active, self-starter with strong technical, analytical and interpersonal skills.

• A team player with good verbal and written skills, capable of working with a team of Architects, Developers, Business/Data Analysts, QA and client stakeholders

• Versatile resource with balanced development skills and business acumen to operate at a fast and accurate speed

• Proficient understanding of distributed computing principles. Continuously evaluate new technologies, innovate and deliver solution for business critical applications

 

Desired Skills

• Object-oriented programming and design experience.

• Degree in Computer Science or equivalent.

• Experience with automated testing methodologies and frameworks, including JUnit, is a plus

• Python IDEs(Django, Flask), data wrangling and analytics in a python-based environment

• Fundamentals of Python - Data Structures, Collections, Pandas for file and other type of data handling, visualizations etc.

• Visual Analytics Tools knowledge ( Tableau )

• Experience with Big Data Analytics & Business Intelligence and Industry standard tools integrated with Hadoop ecosystem. ( R , Python )

• Data Integration, Data Security on Hadoop ecosystem. ( Kerberos )

• Any Big Data certification(ex. Cloudera's CCP, CCA) is a plus

Education

Bachelor's degree