Description

Responsibilities:

• Develop, maintain, and optimize data pipelines to extract, transform, and load large datasets from diverse sources into our data ecosystem.

• Design and implement efficient and scalable data models that align with business requirements, ensuring data integrity and performance.

• Collaborate with cross-functional teams to understand data needs and deliver solutions that meet those requirements.

• Work closely with data scientists, analysts, and software engineers to ensure seamless integration of data solutions into larger systems.

• Identify and resolve data quality issues, ensuring accuracy, reliability, and consistency of the data infrastructure.

• Continuously monitor and improve data pipelines and processes, identifying opportunities for automation and optimization.

• Stay updated with emerging trends, technologies, and best practices in data engineering, data modeling, and backend Java engineering.

• Provide technical guidance and mentorship to junior team members, fostering their growth and development.

 

Requirements:

• Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.

• 5+years of hands-on experience as a Data Engineer, working on complex data projects and implementing data modeling solutions.

Must have:

• Solid understanding of SQL and expertise in working with relational databases (e.g., PostgreSQL, MySQL).

• In-depth knowledge of data modeling techniques and experience with data modeling tools

• Proficiency in designing and optimizing data pipelines using ETL/ELT frameworks and tools (e.g., Informatica, Apache Spark, Airflow, AWS Glue).

• Working knowledge on Data warehousing

• Familiarity with cloud-based data platforms and services (e.g., Snowflake, AWS, Google Cloud, Azure).

• Experience with version control systems (e.g., Git) and agile software development methodologies.

• Strong communication skills to effectively convey technical concepts to both technical and non-technical stakeholders.

• Excellent problem-solving skills and the ability to work independently and collaboratively in a fast-paced environment.

 

Good to Have:

• JAVA 8, REST APIs, and microservices, Spring Boot framework

• UNIX scripting

 

Primary Skillset: Data Engineering

Solid understanding of SQL and expertise in working with relational databases (e.g., DB2, MySQL).

Data Modelling knowledge

cloud-based data platforms like Snowflake

Working knowledge on Data warehousing

Cloud-based data platforms and services (e.g., AWS, Google Cloud, Azure)

Alteryx (good to have)

ETL/ELT tools like Informatica, Apache Spark

UNIX scripting

Education

Bachelor's degree