Minimizing GPU Cost for Deep Learning on Kubernetes
Offered By: Linux Foundation via YouTube
Course Description
Overview
Explore a GPU sharing solution for native Kubernetes to minimize costs and improve efficiency in deep learning tasks. Learn how to define GPU sharing API, implement scheduling without modifying core scheduler code, and integrate GPU isolation with Kubernetes. Discover techniques to run multiple TensorFlow jobs on a single GPU device within a Kubernetes cluster, significantly enhancing GPU usage for AI model development, debugging, and inference services. Gain insights from Alibaba experts on addressing the challenge of wasted GPU resources in clusters and optimizing deep learning workflows on Kubernetes.
Syllabus
Minimizing GPU Cost for Your Deep Learning on Kubernetes - Kai Zhang & Yang Che, Alibaba
Taught by
Linux Foundation
Tags
Related Courses
Introduction to Cloud Infrastructure TechnologiesLinux Foundation via edX Scalable Microservices with Kubernetes
Google via Udacity Google Cloud Fundamentals: Core Infrastructure
Google via Coursera Introduction to Kubernetes
Linux Foundation via edX Fundamentals of Containers, Kubernetes, and Red Hat OpenShift
Red Hat via edX