Chinchilla Explained - Compute-Optimal Massive Language Models
Offered By: Edan Meyer via YouTube
Course Description
Overview
Explore the groundbreaking Chinchilla language model in this 33-minute video lecture. Delve into DeepMind's innovative approach to scaling large language models in a compute-optimal manner, resulting in Chinchilla's superior performance over GPT-3, Gopher, and Megatron-Turing NLG with only 70 billion parameters. Learn about the extensive research involving 400 large models to determine the optimal ratio of parameters and training data. Gain insights into the paper's introduction, methodology, scaling implications, and Chinchilla's overview and performance. Conclude with a summary and critical analysis of this significant advancement in natural language processing.
Syllabus
- Overview
- Paper Intro
- Methods
- Scaling Implications
- Chinchilla Overview
- Chinchilla Performance
- Summary
- Thoughts & Critiques
Taught by
Edan Meyer
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Artificial Intelligence for Robotics
Stanford University via Udacity Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent