Description

Roles and Responsibilities :

Must have Azure, Databricks, Spark, Pyspark, Python experience.
Design and build highly scalable data pipelines for near real-time and batch data ingestion, processing, and data integration.
Coordinates the creation and documentation of ETL pipelines/testing to meet customer needs for reporting, BI solutions, Data Science.
Hands-on experience in configuring Azure Event Hub, Event Grid, Stream Analytics, logic/function apps and JSON.
Expert level skills in Python, Databricks, Azure Data Factory
Be a hands-on mentor and advocate to ensure successful adoption of new tools, processes, and best practices across the organization.
Recognize potential issues and risks during the project implementation and suggest mitigation strategies.
Communicate and own the process of manipulating and merging large datasets.
Expert and key point of contact between the API teams and the project/functional leads
Work directly with business leadership to understand data requirements; propose and develop solutions that enable effective decision-making and drives business objectives.
Prepare advanced project implementation plans which highlight major milestones and deliverables, leveraging standard methods and work planning tools.
Participate in the preparation of high-quality project deliverables that are valued by the business and present them in such a manner that they are easily understood by project stakeholders.
Required skills:

Working knowledge of message-oriented middleware/streaming data technologies such as Kafka/NiFi, MQ, Azure Event Hub
Must have strong programming skills / experience in C# / .NET, Logic App
Must have strong programming skills / experience in Azure Functions using various protocols /triggers, Git/Github
Experienced in the use of ETL tools and techniques and have knowledge of CI/CD
Experience & Expertise in cloud NoSQL databases, ideally Azure/Azure Data Services/Cosmos DB or equivalent
Knowledge of and experience scaling one or more of the popular data streaming and processing technologies like Kafka, Spark, Spark Streaming, etc.
General knowledge and experience with configuration, load-balancing, auto-scaling, monitoring, networking, and problem-solving in a cloud environment
Demonstrates change management and excellent communication skills.
Degree in Information Systems, Computer Science, Engineering, or related field, or the equivalent combination of education, training and experience

Education

ANY GRADUATE