Responsibilities:
• Develop, maintain, and optimize data pipelines to extract, transform, and load large datasets from diverse sources into our data ecosystem.
• Design and implement efficient and scalable data models that align with business requirements, ensuring data integrity and performance.
• Collaborate with cross-functional teams to understand data needs and deliver solutions that meet those requirements.
• Work closely with data scientists, analysts, and software engineers to ensure seamless integration of data solutions into larger systems.
• Identify and resolve data quality issues, ensuring accuracy, reliability, and consistency of the data infrastructure.
• Continuously monitor and improve data pipelines and processes, identifying opportunities for automation and optimization.
• Stay updated with emerging trends, technologies, and best practices in data engineering, data modeling, and backend Java engineering.
• Provide technical guidance and mentorship to junior team members, fostering their growth and development.
Requirements:
• Bachelor’s or Master’s degree in Computer Science, Engineering, or a related field.
• 5+years of hands-on experience as a Data Engineer, working on complex data projects and implementing data modeling solutions.
Must have:
• Solid understanding of SQL and expertise in working with relational databases (e.g., PostgreSQL, MySQL).
• In-depth knowledge of data modeling techniques and experience with data modeling tools
• Proficiency in designing and optimizing data pipelines using ETL/ELT frameworks and tools (e.g., Informatica, Apache Spark, Airflow, AWS Glue).
• Working knowledge on Data warehousing
• Familiarity with cloud-based data platforms and services (e.g., Snowflake, AWS, Google Cloud, Azure).
• Experience with version control systems (e.g., Git) and agile software development methodologies.
• Strong communication skills to effectively convey technical concepts to both technical and non-technical stakeholders.
• Excellent problem-solving skills and the ability to work independently and collaboratively in a fast-paced environment.
Good to Have:
• JAVA 8, REST APIs, and microservices, Spring Boot framework
• UNIX scripting
Primary Skillset: Data Engineering
Solid understanding of SQL and expertise in working with relational databases (e.g., DB2, MySQL).
Data Modelling knowledge
cloud-based data platforms like Snowflake
Working knowledge on Data warehousing
Cloud-based data platforms and services (e.g., AWS, Google Cloud, Azure)
Alteryx (good to have)
ETL/ELT tools like Informatica, Apache Spark
UNIX scripting
Bachelor's degree