YoVDO

Scaling AI Workloads with Kubernetes - Sharing GPU Resources Across Multiple Containers

Offered By: Linux Foundation via YouTube

Tags

Kubernetes Courses Scalability Courses Container Orchestration Courses Containerization Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore how to efficiently scale AI workloads using Kubernetes by sharing GPU resources across multiple containers in this informative conference talk. Delve into the challenges of GPU resource management and learn various techniques for optimizing GPU usage. Discover how to set resource limits to ensure fair and efficient allocation of GPU resources among containers. Gain a solid understanding of leveraging Kubernetes and the NVIDIA device plugin to maximize GPU investments and achieve faster, more accurate results in AI applications. By the end of the talk, acquire valuable insights into overcoming GPU resource bottlenecks and efficiently serving AI workloads in a containerized environment.

Syllabus

Scaling AI Workloads with Kubernetes: Sharing GPU Resources Across Multiple Containers - Jack Ong


Taught by

Linux Foundation

Tags

Related Courses

Introduction to Kubernetes
Linux Foundation via edX
Elastic Cloud Infrastructure: Containers and Services en EspaƱol
Google Cloud via Coursera
Architecting with Google Kubernetes Engine en Portuguese
Google Cloud via Coursera
Building Containerized Applications on AWS
Amazon Web Services via Coursera
Introduction to Containers, Kubernetes and OpenShift
IBM via edX