GPU Scheduling in Kubernetes

Master GPU resource management for AI and ML workloads on Kubernetes. Learn device plugins, time-slicing, Multi-Instance GPU (MIG), scheduling policies, and production best practices for maximizing GPU utilization.

6
Lessons
Hands-On Projects
🕑
Self-Paced
100%
Free

Your Learning Path

Follow these lessons in order, or jump to any topic that interests you.

What You'll Learn

By the end of this course, you'll be able to:

💻

Configure GPU Resources

Set up device plugins, resource requests, and limits to expose and allocate GPUs to Kubernetes pods effectively.

Maximize Utilization

Use time-slicing and MIG to share GPUs across workloads, reducing costs while maintaining performance guarantees.

🚀

Optimize Scheduling

Apply advanced scheduling policies including affinity rules, topology awareness, and priority-based preemption.

📊

Monitor & Scale

Implement GPU monitoring with DCGM, set up autoscaling, and manage multi-tenant GPU clusters in production.