YoVDO

Deploying LLM Workloads on Kubernetes with WasmEdge and Kuasar

Offered By: CNCF [Cloud Native Computing Foundation] via YouTube

Tags

Kubernetes Courses LLM (Large Language Model) Courses Scalability Courses WasmEdge Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the deployment of Large Language Model (LLM) workloads on Kubernetes using WasmEdge and Kuasar in this keynote presentation. Discover how these innovative solutions address challenges in running LLMs, including complex package installations, GPU compatibility issues, scaling limitations, and security vulnerabilities. Learn how WasmEdge enables the development of fast, agile, resource-efficient, and secure LLM applications, while Kuasar facilitates running applications on Kubernetes with faster container startup and reduced management overhead. Witness a demonstration of running Llama3-8B on a Kubernetes cluster using WasmEdge and Kuasar as container runtimes. Gain insights into how Kubernetes enhances efficiency, scalability, and stability in LLM deployment and operations, making this 14-minute presentation essential for those interested in advanced cloud-native AI solutions.

Syllabus

Keynote: Deploying LLM Workloads on Kubernetes by WasmEdge and Kuasar - Tianyang Zhang & Vivian Hu


Taught by

CNCF [Cloud Native Computing Foundation]

Related Courses

Google BARD and ChatGPT AI for Increased Productivity
Udemy
Bringing LLM to the Enterprise - Training From Scratch or Just Fine-Tune With Cerebras-GPT
Prodramp via YouTube
Generative AI and Long-Term Memory for LLMs
James Briggs via YouTube
Extractive Q&A With Haystack and FastAPI in Python
James Briggs via YouTube
OpenAssistant First Models Are Here! - Open-Source ChatGPT
Yannic Kilcher via YouTube