This is a cache of https://developer.ibm.com/series/gpu-mig/. It is a snapshot of the page as it appeared on 2025-11-30T05:41:14.900+0000.
Optimizing GPU resources with NVIDIA Multi-Instance GPU (MIG) in Red Hat OpenShift AI and Kubernetes - IBM Developer

Series

Optimizing GPU resources with NVIDIA Multi-Instance GPU (MIG) in Red Hat OpenShift AI and Kubernetes

Discover how NVIDIA Multi-Instance GPU (MIG) technology enables more efficient utilization of GPUs resources for AI and ML workloads in Red Hat OpenShift AI and Kubernetes.

By

Lorenzo Carleo,

Sarath Chandra Vidya Sagar Machupalli,

Kuan Feng

The articles in this series represent a comprehensive guide of key challenges in AI infrastructure and more importantly, how to optimize GPU resources for AI and ML workloads in containerized environments.

This article series describes key challenges in AI infrastructure: optimizing resource usage (particularly GPUs), reducing costs, and enabling scalability. Also, this series introduces NVIDIA Multi-Instance GPU (MIG) technology, which enables more efficient utilization of GPUs resources for AI and ML workloads in Red Hat OpenShift AI and Kubernetes.

Articles in this series