MiniLLM: Knowledge Distillation of Large Language Models
Offered By: Unify via YouTube
Course Description
Overview
Explore a 52-minute presentation on Knowledge Distillation of Large Language Models by Yuxian Gu, a PhD student at Tsinghua University. Delve into a novel method replacing forward Kullback-Leibler divergence with reverse KLD in standard knowledge distillation approaches for large language models. Discover how this technique prevents student models from overestimating low-probability regions of teacher distributions, resulting in MiniLLMs that generate more precise and higher-quality responses compared to traditional knowledge distillation baselines. Learn about the research paper, its authors, and related resources. Gain insights into AI optimization, language models, and cutting-edge knowledge distillation techniques. Access additional materials including AI research trends, deployment strategies, and connect with the Unify community through various platforms.
Syllabus
MiniLLM: Knowledge Distillation of Large Language Models
Taught by
Unify
Related Courses
Neural Networks for Machine LearningUniversity of Toronto via Coursera Good Brain, Bad Brain: Basics
University of Birmingham via FutureLearn Statistical Learning with R
Stanford University via edX Machine Learning 1—Supervised Learning
Brown University via Udacity Fundamentals of Neuroscience, Part 2: Neurons and Networks
Harvard University via edX