Description

This role requires a candidate with a strong technical background in data engineering (with a focus on data infrastructure and management), software development, and a proactive approach to improving data workflows and quality. Look for candidates with a blend of technical expertise and the ability to collaborate effectively within a team environment. Candidates with a minimum of 3 years of experience as a Data Engineer or equivalent are preferred.

 

Key Responsibilities:

Data Architecture and Design:

  • Work closely with core data users to understand product needs.
  • Design data structures that align with product requirements.
  • Develop, tune, and improve core data assets.

Data Pipeline Development:

  • Build and schedule data transformation pipelines.
  • Automate the lifecycle of data sets, including schema development, deprecation, and iteration.
  • Enhance end-to-end workflow for data users.

Data Quality and Transparency:

  • Define data requirements to ensure high data quality.
  • Identify and implement data observability tooling (lineage, sources, transformations).
  • Improve transparency in data pipelines.

Collaboration and Standards:

  • Collaborate with core team members to develop, test, deploy, and operate high-quality, scalable software.
  • Contribute to raising engineering standards within the team.

 

Required Skillset:

 

Data Modeling and Architecture:

  • Proficiency in designing effective data structures.
  • Experience in architecting scalable and efficient data systems.

Data Pipeline Development:

  • Strong skills in building and optimizing data transformation pipelines.
  • Knowledge of scheduling and automating data workflows.

Data Quality Assurance:

  • Ability to define and implement processes for ensuring data quality.
  • Familiarity with data observability tools and practices.

Software Development:

  • Experience in developing, testing, and deploying scalable software.
  • Collaboration skills to work effectively within a core team.

Automation and Lifecycle Management:

  • Automation skills for managing the lifecycle of data sets.
  • Understanding of schema development, deprecation, and iteration processes.

Communication and Collaboration:

  • Strong communication skills to work closely with data users and core team members.
  • Ability to contribute to and elevate engineering standards.
  • Willingness to stay updated on industry trends and adopt new technologies.

 

For this role, candidates need to work in the PST time zone (04:30 PM - 1:30 AM) and Work from Office (Noida Subject to availability)

Key Skills

Python Snowflake Data Structures Data Modeling and Architecture Data Pipelines Data Warehousing Cloud Infrastructure

Education

Any Graduate