MLOps / LLMOps Engineer

New
Remote Workable locations: IndiaFull-TimeSenior
Salary not disclosed
Apply NowOpens the employer's application page

Job Details

Experience
3–5 years of experience
Required Skills
AWSPythonSQLMLFlowAzureDatabricksGitHub ActionsPrompt EngineeringMLOps

Requirements

  • 3–5 years of experience in MLOps, LLMOps, or ML platform engineering roles.
  • Hands-on experience with Databricks, Delta Lake, Unity Catalog, and ML deployment workflows.
  • Strong experience with CI/CD pipelines using GitHub Actions and infrastructure automation.
  • Experience implementing data quality validation, schema governance, and data contracts.
  • Experience building production-grade ML pipelines with monitoring and observability.
  • Strong security knowledge including RBAC, encryption, data residency, and governance practices.
  • Proficiency in Python, SQL, and distributed data processing frameworks.
  • Experience with LLM pipelines, prompt engineering, RAG workflows, and model optimization.
  • Experience with vector databases, model serving, and MLflow.
  • Experience with Azure and AWS cloud platforms, including security and networking.
  • Bachelor’s or master’s degree in computer science, Software Engineering, or a related field, or equivalent professional experience.

Responsibilities

  • Operationalize model training, evaluation, packaging, and deployment using Databricks, Delta Lake, and medallion architecture.
  • Implement Unity Catalog model governance, lineage tracking, and access control.
  • Develop reusable job templates, cluster policies, and standardized deployment patterns.
  • Deploy and manage ML and GenAI solutions including risk scoring, anomaly detection, predictive maintenance, NLP, and RAG pipelines.
  • Build and optimize LLM pipelines using vector databases, model serving endpoints, and inference workflows.
  • Optimize models using quantization, caching, and performance tuning techniques.
  • Implement batch and real-time inference pipelines with defined SLAs.
  • Implement data contracts, schema validation, and data quality checks across ML pipelines.
  • Ensure secure handling of sensitive data including PII detection, classification, and obfuscation.
  • Maintain full lineage from data sources to deployed models and serving endpoints.
  • Enforce data residency, governance, and compliance policies.
  • Implement CI/CD pipelines using GitHub Actions and Databricks Asset Bundles.
  • Automate deployments across DEV, QA, and PROD environments.
  • Develop unit and integration tests for data pipelines and ML models.
  • Ensure version control, reproducibility, and automated deployment workflows.
  • Monitor pipeline health, model performance, drift, and system reliability.
  • Implement alerting, incident response workflows, and automated ticketing.
  • Track LLM performance metrics including latency, hallucination rates, and API costs.
  • Develop runbooks, disaster recovery procedures, and operational documentation.
  • Apply tagging policies and cost tracking for ML infrastructure.
  • Support budget monitoring, cost optimization, and resource management.
View Full Description & ApplyYou'll be redirected to the employer's site
View details
Apply Now