Responsibilities:
π Design, develop, and implement data pipelines using cutting-edge tools such as Apache Airflow.
βοΈ Monitor and optimize ELT data pipelines for peak performance, reliability, and scalability.
π Troubleshoot and resolve issues related to data pipelines, ensuring data integrity and quality.
π‘ Utilize dbt for data transformation, optimizing its usage and operations while establishing standards and best practices.
π Implement and maintain data governance practices, including data lineage and metadata management.
π₯ Collaborate closely with the DevOps team to ensure seamless integration and deployment of data pipelines using GitLab.
π° Stay up-to-date with industry trends and best practices in data engineering and orchestration.
Requirements:
π Proven experience as a DataOps Engineer or in a similar role with a minimum of 10 years of hands-on experience.
π Strong proficiency in data orchestration tools, particularly Apache Airflow.
π‘ Solid understanding of data engineering concepts and best practices.
π» Proficiency in Git, Python, and SQL.
π Experience with data transformation tools, especially dbt.
βοΈ Experience with AWS cloud platform (Redshift, MWAA, Glue, Lambda, S3).
π Familiarity with data quality and data observability tools.
π Familiarity with DevOps tools (e.g., GitLab) and concepts (CI/CD, containerization).
π€ Strong problem-solving and troubleshooting skills.
π£ Excellent communication and collaboration abilities
Bachelor's degree in Computer Science