Bringing LLMs Everywhere Through Machine Learning Compilation
Offered By: The ASF via YouTube
Course Description
Overview
Explore the groundbreaking MLC-LLM project, an open-source initiative based on Apache TVM that enables running large language models (LLMs) on various devices, including PCs, mobile devices, and WebGPU with GPU acceleration. Delve into the challenges of deploying computationally intensive LLMs beyond traditional server environments with cloud GPUs. Learn how machine learning compilation techniques are revolutionizing the accessibility of generative AI and LLMs, potentially transforming numerous domains by bringing these powerful models to a wider range of devices and platforms.
Syllabus
Bringing Llm To Everywhere Via Machine Learning Compilation
Taught by
The ASF
Related Courses
WebGPU Tutorial - Advanced Graphics on the Web CoursefreeCodeCamp WebGPU: Modern GPU Rendering and Compute for Web and Android
Android Developers via YouTube Throw Away Your Xbox - The Future of Games Is Written in JavaScript
NDC Conferences via YouTube Rust for Medical Visualization - May 2021
Rust via YouTube Portable Graphics Abstraction in Rust with gfx-rs
Rust via YouTube