YoVDO

Scaling AI Workloads with Kubernetes - Sharing GPU Resources Across Multiple Containers

Offered By: Linux Foundation via YouTube

Tags

Kubernetes Courses Scalability Courses Container Orchestration Courses Containerization Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore how to efficiently scale AI workloads using Kubernetes by sharing GPU resources across multiple containers in this informative conference talk. Delve into the challenges of GPU resource management and learn various techniques for optimizing GPU usage. Discover how to set resource limits to ensure fair and efficient allocation of GPU resources among containers. Gain a solid understanding of leveraging Kubernetes and the NVIDIA device plugin to maximize GPU investments and achieve faster, more accurate results in AI applications. By the end of the talk, acquire valuable insights into overcoming GPU resource bottlenecks and efficiently serving AI workloads in a containerized environment.

Syllabus

Scaling AI Workloads with Kubernetes: Sharing GPU Resources Across Multiple Containers - Jack Ong


Taught by

Linux Foundation

Tags

Related Courses

Fundamentals of Containers, Kubernetes, and Red Hat OpenShift
Red Hat via edX
Configuration Management for Containerized Delivery
Microsoft via edX
Getting Started with Google Kubernetes Engine - Español
Google Cloud via Coursera
Getting Started with Google Kubernetes Engine - 日本語版
Google Cloud via Coursera
Architecting with Google Kubernetes Engine: Foundations en Español
Google Cloud via Coursera