Skills and Experience
Bachelor's Degree Computer Science, Computer Engineering, Information Technology, Software Engineering or equivalent technical discipline and 10+ years of experience in software engineering with a strong background in DevOps and Infrastructure as Code, supporting Machine Learning and Data Science workloads preferred. or
Master's Degree Computer Science, Computer Engineering, Information Technology, Software Engineering or equivalent technical discipline and 5+ years of experience in software engineering with a strong background in DevOps and Infrastructure as Code, supporting Machine Learning and Data Science workloads preferred.
Expertise on code versioning tools, such as Gitlab, GitHub, Azure DevOps, Bitbucket etc., GitHub Preferred, familiar with branch level code repository management.
Experience deploying Machine Learning solutions on cloud platforms (e.g., AWS, Azure, or GCP). Databricks, and AWS Preferred.
Proficient with GitHub actions to automate testing and deployment of data and Client workloads from CI/CD provider to Databricks.
Strong knowledge of infrastructure automation tools such as Terraform, Ansible, CloudFormation etc.
Experience with data processing frameworks/tools/platform such as Databricks, Apache Spark, Kafka, Flink, AWS cloud services for batch processing, batch streaming and streaming.
Experience containerizing analytical models using Docker and Kubernetes or other container orchestration platforms.
Technical expertise across all deployment models on public cloud, private cloud, and on-premises infrastructure.
Responsibilities
Develop automated build and deployment processes to enable continuous delivery of software releases, enhance the existing CI/CD pipelines for AIML application development and deployment.
Collaborate with data scientists, data engineers, data analysts, software engineers, IT specialists, and stakeholders to accelerate deployment of AI applications via CI/CD pipelines and maintain the SLAs of those applications at the centralized platform.
Design, develop and maintain infrastructure using infrastructure as code tools such as Terraform, Ansible, CloudFormation etc.
Templatize existing Databricks CLI codes to manage Databricks platform as code for AIML data pipelines (batch processing, batch streaming and streaming) and model serving endpoints.
Enhance the existing DevOps practices to improve the overall AIML application development lifecycle.
Work closely with cross-functional teams to ensure that applications are highly available and scalable.
Collaborate with development teams and cloud platform team to ensure that infrastructure meets the requirements of the application.
Establish and maintain best practices for cloud security, compliance, and cost optimization.
Bachelor's degree in Computer Science