Fine-tuning Llama 3 on Wikipedia Datasets for Low-Resource Languages
Offered By: Trelis Research via YouTube
Course Description
Overview
Explore the process of fine-tuning Llama 3 for low-resource languages using Wikipedia datasets in this comprehensive 44-minute tutorial. Learn how to create a HuggingFace dataset using WikiExtractor, set up Llama 3 fine-tuning with LoRA, and implement dataset blending to prevent catastrophic forgetting. Dive into trainer setup, parameter selection, and loss inspection. Gain insights on learning rates, annealing, and additional tips for improving your fine-tuning results. Access provided resources including slides, dataset links, and code repositories to enhance your learning experience.
Syllabus
Fine-tuning Llama 3 for a low resource language
Overview of Wikipedia Dataset and Loss Curves
Video overview
HuggingFace Dataset creation with WikiExtractor
Llama 3 fine-tuning setup, incl. LoRA
Dataset blending to avoid catastrophic forgetting
Trainer setup and parameter selection
Inspection of losses and results
Learning Rates and Annealing
Further tips and improvements
Taught by
Trelis Research
Related Courses
Microsoft Bot Framework and Conversation as a PlatformMicrosoft via edX Unlocking the Power of OpenAI for Startups - Microsoft for Startups
Microsoft via YouTube Improving Customer Experiences with Speech to Text and Text to Speech
Microsoft via YouTube Stanford Seminar - Deep Learning in Speech Recognition
Stanford University via YouTube Select Topics in Python: Natural Language Processing
Codio via Coursera