AI for DevOps & Platform Engineers
MLOps is DevOps for AI. If you know CI/CD, you're closer than you think.
The AI era creates huge new infrastructure work — model deployment, GPU management, drift monitoring, and AI observability. This track teaches DevOps engineers to build and operate production ML infrastructure using patterns they already understand.
Curriculum
MLOps Foundations for DevOps
How ML workflows differ from software. Experiment tracking, model versioning, the MLOps lifecycle — mapped to DevOps concepts you already know.
Containerising & Serving ML Models
Docker for ML, model artifacts, dependency management, FastAPI, Triton Inference Server, and vLLM for LLMs.
CI/CD for ML Models
Automated ML testing, model validation gates, canary/blue-green/shadow deployments, GitHub Actions integration.
GPU Infrastructure & Cost Management
GPU instance selection, spot instances, inference optimisation, quantisation, and cost dashboards.
Model Monitoring & Drift Detection
Data drift, concept drift, performance monitoring, anomaly detection, alerting pipelines.
Feature Stores & Data Pipelines
Feature engineering, feature stores (Feast, Tecton), data versioning with DVC.
AI Observability & Incidents
Token cost dashboards, latency profiling, prompt logging, AI incident runbooks.
Production AI Architecture
Reference architectures for RAG, agents, batch inference, real-time scoring — from HYVE's UAE deployments.