Description

Your Impact:

Data Ingestion, Integration and Transformation
Data Storage and Computation Frameworks, Performance Optimizations
Analytics & Visualizations
Infrastructure & Cloud Computing
Data Management Platforms
Build functionality for data ingestion from multiple heterogeneous sources in batch & real-time
Build functionality for data analytics, search and aggregation
Qualifications

 

Your Skills & Experience:

Bachelor’s degree and year of work experience of 6 to 8 years or any combination of education, training, and/or experience that demonstrates the ability to perform the duties of the position
Minimum 3 years of experience in Big Data technologies
Hands-on experience with the Hadoop stack – HDFS, sqoop, Kafka, Pulsar, NiFi, Spark, Spark Streaming, Flink, Storm, hive, oozie, airflow, and other components required in building end-to-end data pipelines. Working knowledge of real-time data pipelines is added advantage.
Strong experience in at least the programming language Java, Scala, and Python. Python preferable
Hands-on working knowledge of NoSQL and MPP data platforms like Hbase, MongoDB, Cassandra, AWS Redshift, Azure SQLDW, GCP BigQuery, etc.
Well-versed and working knowledge with data platform-related services on GCP
Set Yourself Apart With:

Good knowledge of traditional ETL tools (Informatica, Talend, etc) and database technologies (Oracle, MySQL, SQL Server, Postgres) with hands-on experience
Knowledge of data governance processes (security, lineage, catalog) and tools like Collibra, Alation, etc
Knowledge of distributed messaging frameworks like ActiveMQ / RabbiMQ / Solace, search & indexing, and Microservices architectures
Performance tuning and optimization of data pipelines
Cloud data specialty and other related Big data technology certifications

Education

Any Graduate