Required Skills
• 3-6years experience in Hadoop stack and storage technologies, HDFS, MapReduce, Yarn, HIVE, sqoop, Impala , spark, flume, kafka and oozie
• Extensive Knowledge on Bigdata Enterprise architecture (Cloudera preferred)
• Excellent analytical capabilities - Strong interest in algorithms
• Experienced in HBase, RDBMS, SQL, ETL and data analysis
• Experience in No SQL Technologies (ex., Cassandra/ MongoDB, etc )
• Experienced in scripting(Unix/Linux) and scheduling (Autosys)
• Experience with team delivery/release processes and cadence pertaining to code deployment and release
• Research oriented, motivated, pro-active, self-starter with strong technical, analytical and interpersonal skills.
• A team player with good verbal and written skills, capable of working with a team of Architects, Developers, Business/Data Analysts, QA and client stakeholders
• Versatile resource with balanced development skills and business acumen to operate at a fast and accurate speed
• Proficient understanding of distributed computing principles. Continuously evaluate new technologies, innovate and deliver solution for business critical applications
Desired Skills
• Object-oriented programming and design experience.
• Degree in Computer Science or equivalent.
• Experience with automated testing methodologies and frameworks, including JUnit, is a plus
• Python IDEs(Django, Flask), data wrangling and analytics in a python-based environment
• Fundamentals of Python - Data Structures, Collections, Pandas for file and other type of data handling, visualizations etc.
• Visual Analytics Tools knowledge ( Tableau )
• Experience with Big Data Analytics & Business Intelligence and Industry standard tools integrated with Hadoop ecosystem. ( R , Python )
• Data Integration, Data Security on Hadoop ecosystem. ( Kerberos )
• Any Big Data certification(ex. Cloudera's CCP, CCA) is a plus
Bachelor's degree