YoVDO

Accelerator Chaining for Efficient AI/ML Workloads in Kubernetes

Offered By: CNCF [Cloud Native Computing Foundation] via YouTube

Tags

Kubernetes Courses FPGA Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore how to efficiently handle large AI/ML workloads in Kubernetes using accelerator chaining in this 37-minute conference talk from CNCF. Learn about extending Kubernetes with Custom Resource-based architecture and operators to orchestrate and configure device chains like FPGAs, GPUs, TPUs, and ASICs. Discover the benefits of direct data transfer between devices, including reduced memory copies, decreased CPU overheads, and lower latency. Gain insights into deploying these workloads easily and understand future developments with Dynamic Resource Allocation (DRA) support and CNI extensions. Presented by Sampath Priyankara from Nippon Telegraph and Telephone Corporation and Masataka Sonoda from Fujitsu Limited, this talk offers valuable knowledge for optimizing AI/ML performance in Kubernetes environments.

Syllabus

Accelerators(FPGA/GPU) Chaining to Efficiently Handle Large AI/ML Workloads in K8s


Taught by

CNCF [Cloud Native Computing Foundation]

Related Courses

Introduction to Cloud Infrastructure Technologies
Linux Foundation via edX
Scalable Microservices with Kubernetes
Google via Udacity
Google Cloud Fundamentals: Core Infrastructure
Google via Coursera
Introduction to Kubernetes
Linux Foundation via edX
Fundamentals of Containers, Kubernetes, and Red Hat OpenShift
Red Hat via edX