Chat Fine-Tuning for LLMs - Instruction Format, Datasets, and Implementation
Offered By: Trelis Research via YouTube
Course Description
Overview
Explore the intricacies of chat fine-tuning in this 20-minute video from Trelis Research. Dive into the world of LLM instruction fine-tuning, understanding its importance and the technicalities involved. Learn about special tokens, stop tokens, and various prompt formats including Guanaco, Llama 2, and chatml. Discover instruction fine-tuning datasets and witness a practical demonstration of fine-tuning a model for chat using Jupyter lab. Gain insights from the results and benefit from pro tips shared by experts. Access additional resources including presentation slides, a Runpod affiliate link, and a Llama 2 instruction fine-tuning dataset. For those seeking advanced fine-tuning capabilities, explore the option to purchase access to comprehensive scripts and notebooks for unsupervised and supervised fine-tuning, dataset preparation, embeddings creation, and quantization.
Syllabus
LLM instruction fine-tuning
Why do instruction fine-tuning?
Understanding special tokens and stop tokens
Instruction fine-tuning format
Guanaco instruction fine-tuning
Llama 2 prompt format
chatml prompt format
Instruction fine-tuning datasets
Fine-tuning a model for chat in Jupyter lab
Chat fine-tuning results
Pro tips
Taught by
Trelis Research
Related Courses
Introduction to Data Science in PythonUniversity of Michigan via Coursera Julia Scientific Programming
University of Cape Town via Coursera Python for Data Science
University of California, San Diego via edX Probability and Statistics in Data Science using Python
University of California, San Diego via edX Introduction to Python: Fundamentals
Microsoft via edX