YoVDO

Leveraging Wasm for Portable AI Inference Across GPUs, CPUs, OS and Cloud-Native Environments

Offered By: CNCF [Cloud Native Computing Foundation] via YouTube

Tags

WebAssembly Courses Kubernetes Courses GPU Computing Courses Edge Computing Courses Cross-Platform Development Courses Cloud Native Computing Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the advantages of using WebAssembly (Wasm) for AI inference tasks in cloud-native ecosystems in this 35-minute conference talk. Learn how Wasm enables developers to create applications on their personal computers that can be uniformly executed across various hardware platforms, including GPUs and CPUs, operating systems, and edge cloud environments. Discover how Wasm and Wasm runtime facilitate seamless integration into cloud-native frameworks, enhancing the deployment and scalability of AI applications. Gain insights into how Wasm provides a flexible and efficient solution for diverse cloud-native architectures, including Kubernetes, allowing developers to fully harness the potential of large language models, particularly open-source ones. Understand how leveraging Wasm's cross-platform capabilities can maximize the potential of AI applications, ensuring consistency, cost-effectiveness, and efficiency in AI inference across different computing environments.

Syllabus

Leveraging Wasm for Portable AI Inference Across GPUs, CPUs, OS & Cloud...- Miley Fu & Hung-Ying Tai


Taught by

CNCF [Cloud Native Computing Foundation]

Related Courses

Моделирование биологических молекул на GPU (Biomolecular modeling on GPU)
Moscow Institute of Physics and Technology via Coursera
Practical Deep Learning For Coders
fast.ai via Independent
GPU Architectures And Programming
Indian Institute of Technology, Kharagpur via Swayam
Perform Real-Time Object Detection with YOLOv3
Coursera Project Network via Coursera
Getting Started with PyTorch
Coursera Project Network via Coursera