Description

We are looking for an experienced Data Architect (Azure) with the below skill set.

Employment Type: Long Term Contract
IMMIDIATE JOINERS ONLY

Location: Location- Mumbai/Bangalore (Work from Home)

Work Timings: IST timings


Role & Responsibilities:

 

  • Work with cloud engineers and customers to solve big data problems by developing utilities for migration, storage and processing on Azure Cloud.
  • Design and build a cloud migration strategy for cloud and on-premise applications.
  • Diagnose and troubleshoot complex distributed systems problems and develop solutions with a significant impact at massive scale.
  • Build tools to ingest and jobs to process several terabytes or petabytes per day.
  • Design and develop next-gen storage and compute solutions for several large customers.
  • Define Data Architecture for the Data Science teams and participate in review and walk-through sessions for model fit and model productionization
  • Provide thought leadership on data integrity & quality for data science workloads
    Be involved in proposals, RFPs and provide effort estimates, solution design etc.
  • Communicate with a wide set of teams, including Infrastructure, Network, Engineering, DevOps, SiteOps teams, and cloud customers.
  • Build advanced tooling for automation, testing, monitoring, administration, and data operations across multiple cloud clusters.
  • Better understanding of Data modeling and governance

    Must have:
  • 8+ years' experience of Hands-on in Data Architect (Azure), Data Modelling, SQL Performance Tuning, Azure SQL, Azure SQL DWH, Data Bricks
  • Strong software development skills in at least one of: Python, Java or Scala. SQL commands
  • Experience building and deploying cloud-based solutions at scale.
  • Experience in developing Big Data solutions (migration, storage, processing)
  • Experience building and supporting large-scale systems in a production environment.
  • Designing and development of ETL pipeline
  • Modern Azure data warehouse design skills
  • Requirement gathering and understanding of the problem statement.
  • End-to-end ownership of the entire delivery of the project
  • Designing and documentation of the solution
  • Knowledge of RDBMS & NoSQL databases
  • Any of Kafka, Kinesis, Cloud pub-sub
  • Cloud Platforms Azure (GCP & AWS Good to have)
  • Any of Apache Hadoop/CDH/HDP/EMR/Google DataProc/HD-Insights Distributed processing Frameworks.
  • One or more of MapReduce, Apache Spark, Apache Storm, Apache Flink. Database/warehouse
  • Hive, HBase, and at least one cloud native services Orchestration Frameworks
  • Any of Airflow, Oozie, Apache NiFi, Google Data Flow Message/Event Solutions
  • Reporting tool exposure (at least one of Power BI, Tableau, Looker)
  • Enable best practices w.r.t data handling.

Leadership qualities:

 

  • Provide thought leadership to the team and bring industry best practices to the project.
  • Ability to lead technology teams and provide them mentorship / support to accelerate performance.
  • Ability to handle conflicts effectively by managing internal and external stakeholders
  • Experience in leading multiple large projects as well as a deep understanding of Agil developments
  • Effective communication with all the stakeholders involved.
  • Communicate clearly about complex subjects and technical plans with technical and non- technical audiences.

 

Good to have:

 

  • Experience with Healthcare datasets - EMR/EHR, Claims, Population health data etc.,o Kubernetes or Swarm
  • Terraforms,
  • Container Orchestration
  • Project execution methodologies.

Education

Any Graduate