Fine-tuning LLMs 30x Faster - Unsloth AI's Approach
Offered By: Aleksa Gordić - The AI Epiphany via YouTube
Course Description
Overview
Explore a comprehensive video interview with Daniel Han from Unsloth AI, delving into techniques for accelerating LLM fine-tuning by up to 30 times. Learn about Han's bug-hunting process, the use of Desmos for gradient checking, and an in-depth analysis of Gemma bugs. Discover insights on runtime bottlenecks, comparisons with llm.c, and discussions on multi-GPU support and reproducibility in machine learning research. Gain valuable knowledge on optimizing LLM performance and understanding the intricacies of fine-tuning processes in this informative hour-long session.
Syllabus
00:00:00 - Intro
00:01:40 - Hyperstack GPUs! sponsored
00:03:21 - About Daniel, getting started with ML
00:06:30 - Using Desmos to check grads
00:11:00 - Deep dive into Gemma bugs
00:38:00 - approximate GELU bug
00:50:00 - What are the bottlenecks to speeding up the runtime?
00:54:15 - comparison with llm.c?
00:58:30 - Is multi-GPU coming to Unsloth? :
01:00:00 - Reproducibility in ML research
Taught by
Aleksa Gordić - The AI Epiphany
Related Courses
TensorFlow: Working with NLPLinkedIn Learning Introduction to Video Editing - Video Editing Tutorials
Great Learning via YouTube HuggingFace Crash Course - Sentiment Analysis, Model Hub, Fine Tuning
Python Engineer via YouTube GPT3 and Finetuning the Core Objective Functions - A Deep Dive
David Shapiro ~ AI via YouTube How to Build a Q&A AI in Python - Open-Domain Question-Answering
James Briggs via YouTube