YoVDO

Deploying LLM Workloads on Kubernetes with WasmEdge and Kuasar

Offered By: Linux Foundation via YouTube

Tags

Kubernetes Courses Cloud Computing Courses LLM (Large Language Model) Courses GPU Computing Courses Scalability Courses Containerization Courses WasmEdge Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the deployment of Large Language Model (LLM) workloads on Kubernetes using WasmEdge and Kuasar in this informative keynote presentation. Discover how these innovative technologies address challenges in running LLMs, including complex package installations, GPU compatibility issues, scaling limitations, and security vulnerabilities. Learn about WasmEdge's solution for developing fast, agile, resource-efficient, and secure LLM applications, as well as Kuasar's ability to enable faster container startup and reduced management overhead on Kubernetes. Witness a demonstration of running Llama3-8B on a Kubernetes cluster using WasmEdge and Kuasar as container runtimes. Gain insights into how Kubernetes enhances efficiency, scalability, and stability in LLM deployment and operations, providing valuable knowledge for developers and IT professionals working with advanced AI models.

Syllabus

Keynote: Deploying LLM Workloads on Kubernetes by WasmEdge and Kuasar - Tianyang Zhang & Vivian Hu


Taught by

Linux Foundation

Tags

Related Courses

Fundamentals of Containers, Kubernetes, and Red Hat OpenShift
Red Hat via edX
Configuration Management for Containerized Delivery
Microsoft via edX
Getting Started with Google Kubernetes Engine - Español
Google Cloud via Coursera
Getting Started with Google Kubernetes Engine - 日本語版
Google Cloud via Coursera
Architecting with Google Kubernetes Engine: Foundations en Español
Google Cloud via Coursera