Core Job Responsibilities
Develop end-to-end ML pipelines encompassing the ML lifecycle from data ingestion, data transformation, model training, model validation, model serving, and model evaluation over time.
Collaborate closely with AI scientists to accelerate productionization of ML algorithms.
Setup CI/CD/CT pipelines, model repository for ML algorithms
Deploy models as a service both on-cloud and on-prem.
Learn and apply new tools, technologies, and industry best practices.
Key Qualifications
MS in Computer Science, Software Engineering, or equivalent field
Experience with Cloud Platforms, especially GCP and related skills: Docker, Kubernetes, edge computing
Familiarity with task orchestration tools such as MLflow, Kubeflow, Airflow, Vertex AI, Azure ML, etc.
Fluency in at least one general purpose programming language. Python - Required
Strong skills in the following: Linux/Unix environment, testing, troubleshooting, automation, Git, dependency management, and build tools (GCP Cloud Build, Jenkins, Gitlab CI/CD, Github Actions, etc.).
Data engineering skills are a plus, such as Beam, Spark, Pandas, SQL, Kafka, GCP Dataflow, etc.
5+ years of experience, including academic experience, in any of the above.
Any Graduate