Develop, Build and Deploy Lightweight AI Services with WebAssembly
Offered By: CNCF [Cloud Native Computing Foundation] via YouTube
Course Description
Overview
Explore the development, building, and deployment of lightweight AI services using WebAssembly in this comprehensive tutorial. Learn how to create efficient PyTorch and TensorFlow inference functions, extend large language models like ChatGPT using Rust, and deploy AI inference functions with Kubernetes. Discover the advantages of WebAssembly as a powerful alternative runtime for AI inference workloads, offering lightweight, fast, and cloud-safe solutions. Gain hands-on experience in creating Wasm-based AI services that address the challenges of resource-intensive inference demands, particularly in the context of large language models. Understand the architecture and fundamentals of WebAssembly-based AI services, and explore how they compare to traditional Python-based approaches in terms of size, speed, and resource consumption.
Syllabus
Develop, Build & Deploy Lightweight AI Services with WebAssembly - Kinfey Lo, Vivian Hu, Jianbai Ye
Taught by
CNCF [Cloud Native Computing Foundation]
Related Courses
Introduction to Cloud Infrastructure TechnologiesLinux Foundation via edX Scalable Microservices with Kubernetes
Google via Udacity Google Cloud Fundamentals: Core Infrastructure
Google via Coursera Introduction to Kubernetes
Linux Foundation via edX Fundamentals of Containers, Kubernetes, and Red Hat OpenShift
Red Hat via edX