Chinchilla Explained - Compute-Optimal Massive Language Models
Offered By: Edan Meyer via YouTube
Course Description
Overview
Explore the groundbreaking Chinchilla language model in this 33-minute video lecture. Delve into DeepMind's innovative approach to scaling large language models in a compute-optimal manner, resulting in Chinchilla's superior performance over GPT-3, Gopher, and Megatron-Turing NLG with only 70 billion parameters. Learn about the extensive research involving 400 large models to determine the optimal ratio of parameters and training data. Gain insights into the paper's introduction, methodology, scaling implications, and Chinchilla's overview and performance. Conclude with a summary and critical analysis of this significant advancement in natural language processing.
Syllabus
- Overview
- Paper Intro
- Methods
- Scaling Implications
- Chinchilla Overview
- Chinchilla Performance
- Summary
- Thoughts & Critiques
Taught by
Edan Meyer
Related Courses
Natural Language ProcessingColumbia University via Coursera Natural Language Processing
Stanford University via Coursera Introduction to Natural Language Processing
University of Michigan via Coursera moocTLH: Nuevos retos en las tecnologĂas del lenguaje humano
Universidad de Alicante via MirĂadax Natural Language Processing
Indian Institute of Technology, Kharagpur via Swayam