YoVDO

Efficient and Cross-Platform LLM Inference in the Heterogeneous Cloud

Offered By: Linux Foundation via YouTube

Tags

WebAssembly Courses Artificial Intelligence Courses Cloud Computing Courses Edge Computing Courses Cross-Platform Development Courses Containerization Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the challenges and solutions for running AI/LLM applications in heterogeneous cloud environments through this 33-minute conference talk by Michael Yuan from Second State. Learn about the limitations of traditional Linux containers for AI workloads and discover how cloud-native WebAssembly (Wasm) offers a portable bytecode format that abstracts away hardware differences. Gain insights into the W3C WASI-NN standard and its role in enabling cross-platform LLM applications. Understand how to develop LLM applications in Rust on a Macbook and deploy them seamlessly on Nvidia cloud servers or ARM NPU devices without recompilation. Explore the advantages of Wasm apps in container management using tools like Docker, Podman, and Kubernetes. Dive into WasmEdge's implementation of WASI-NN and its support for various AI/LLM applications. Acquire practical skills for building and running LLM applications across local, edge, and cloud devices using a single binary application.

Syllabus

Efficient and Cross-Platform LLM Inference in the Heterogenous Cloud - Michael Yuan, Second State


Taught by

Linux Foundation

Tags

Related Courses

Game Development for Modern Platforms
Michigan State University via Coursera
Introduction to Xamarin.Forms
Microsoft via edX
Building R Packages
Johns Hopkins University via Coursera
Multiplatform Mobile App Development with NativeScript
The Hong Kong University of Science and Technology via Coursera
Build Mobile Applications with SAP Cloud Platform Mobile Services
SAP Learning