FwdLLM: Efficient Federated Finetuning of Large Language Models with Perturbed Inferences
Offered By: USENIX via YouTube
Course Description
Overview
Explore an innovative approach to federated fine-tuning of Large Language Models (LLMs) in this 20-minute conference talk from USENIX ATC '24. Dive into FwdLLM, a novel protocol designed to enhance efficiency in Federated Learning (FL) for LLMs on mobile devices. Learn how the researchers from Beijing University of Posts and Telecommunications address the challenge of balancing LLM complexity with mobile resource constraints. Discover the key components of FwdFL, including backpropagation-free training methods, adaptive computational load allocation, and discriminative sampling of perturbed predictions. Gain insights into the significant advantages of this approach, such as faster convergence and reduced memory footprint, and understand how it enables federated billion-parameter LLMs on commercial off-the-shelf mobile devices for the first time.
Syllabus
USENIX ATC '24 - FwdLLM: Efficient Federated Finetuning of Large Language Models with Perturbed...
Taught by
USENIX
Related Courses
LLaMA- Open and Efficient Foundation Language Models - Paper ExplainedYannic Kilcher via YouTube Alpaca & LLaMA - Can it Compete with ChatGPT?
Venelin Valkov via YouTube Experimenting with Alpaca & LLaMA
Aladdin Persson via YouTube What's LLaMA? ChatLLaMA? - And Some ChatGPT/InstructGPT
Aladdin Persson via YouTube Llama Index - Step by Step Introduction
echohive via YouTube