Description

Job Description:

Experience Desired: 10+ Years.      

Responsibilities:

Interview process - 3 interview round and 1 client round.

Required Skills/Qualifications:

  • 10-14 years of relevant experience
  • Bachelor's and/or master’s degree in computer science or equivalent experience.
  • Strong communication, analytical and problem-solving skills with a high attention to detail. 

Desired Experience:

  • At least two years of experience building and leading highly complex, technical engineering teams.
  • Strong hands-on experience in Databricks
  • Implement scalable and sustainable data engineering solutions using tools such as Databricks, Azure, Apache Spark, and Python. The data pipelines must be created, maintained, and optimized as workloads move from development to production for specific use cases.
  • Experience managing distributed teams preferred.
  • Comfortable working with ambiguity and multiple stakeholders.
  • Comfortable working cross functionality with product management and directly with customers; ability to deeply understand product and customer personas.
  • Expertise on Azure Cloud platform
  • Good SQL knowledge
  • Knowledge on orchestrating workloads on cloud
  • Ability to set and lead the technical vision while balancing business drivers 
  • Strong experience with PySpark, Python programming
  • Proficiency with APIs, containerization and orchestration is a plus
  • Experience handling large and complex sets of data from various sources and databases
  • Solid grasp of database engineering and design principles.
  • Experience with Unity Catalog.
  • Familiarity with CI/CD methods desired
  • Good to have Teradata Experience (not Mandatory) 

Responsibilities:

  • Manage end to end delivery by Investigating problem areas, working cross-functionally with product manager & other stakeholders
  • Follow the Agile development methodology;  think strategically and execute methodically
  • Develop and manage capacity and growth projection forecasts of the environment within budgets
  • Create and maintain optimal data pipeline architecture, Assemble large, complex data sets that meet functional / non-functional business requirements
  • Drive optimization, testing and tooling to improve quality of solutions
  • Manage teams that build and operate high volume distributed systems in a SaaS environment
  • Great at devising efficient processes that increase velocity and quality
  • Train counterparts in the data pipelining and preparation techniques, which make it easier for them to integrate and consume the data they need for their own use cases.
  • Promote the available data and analytics capabilities and expertise to business unit leaders and educate them in leveraging these capabilities in achieving their business goals.
  • Accountable for operational effectiveness in performance, uptime, release management of the enterprise data platform 


 

Education

Bachelor's degree