Description

You Are Someone Who Can


You will be involved in the design of data solutions using Hadoop based technologies along with Hadoop, Azure, HDInsight for Cloudera based Data Late using Scala Programming.
Liaise and be part of our extensive GCP community, contributing in the knowledge exchange learning programme of the platform.
Be required to showcase your GCP Data engineering experience when communicating with business team on their requirements, turning these into technical data solutions.
Be required to build and deliver Data solutions using GCP products and offerings.
Hands on and deep experience working with Google Data Products (e.g. Big Query, Dataflow, Dataproc, AI Building Blocks, Looker, Cloud Data Fusion, Dataprep, etc.).
Experience in Spark /Scala / Python/Java / Kafka.
Responsible to Ingest data from files, streams and databases. Process the data with Hadoop, Scala, SQL Database, Spark, ML, IoT
Develop programs in Scala and Python as part of data cleaning and processing
Responsible to design and develop distributed, high volume, high velocity multi-threaded event processing systems
Develop efficient software code for multiple use cases leveraging Python and Big Data technologies for various use cases built on the platform
Provide high operational excellence guaranteeing high availability and platform stability
Implement scalable solutions to meet the ever-increasing data volumes, using big data/cloud technologies Pyspark, Kafka, any Cloud computing etc


You Bring In


Minimum 4+ years of experience in Big Data technologies
Good to have experience in Cloud, GCP, AWS, Azure Data Engineering with background in Spark/Python/Scala / Java.
Proficient in any of the programming languages – Python, Scala or Java
Mandatory experience in Mid to Expert Level programming capabilities in a large-scale enterprise
In-depth experience in modern data platform components such as the Hadoop, Hive, Pig, Spark, Python, Scala, etc
Experience with Distributed Versioning Control environments such as GIT
Familiarity with development tools – experience on either IntelliJ / Eclipse / VSCode IDE, Build Tool Maven
Demonstrated experience in modern API platform design including how modern UI’s are built consuming services / APIs.
Experience on Azure cloud including Data Factory, Databricks, Data Lake Storage is highly preferred.
Solid experience in all phases of Software Development Lifecycle – plan, design, develop, test, release, maintain and support, decommission.


Your Qualification Is


B.E/B.Tech/M.C.A/MSc (preferably in Computer Science/IT)

Education

Any Graduate