YoVDO

The Pitfalls of Next-token Prediction in Language Models

Offered By: Simons Institute via YouTube

Tags

Generalization Courses Artificial Intelligence Courses Machine Learning Courses Transformers Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a thought-provoking lecture that delves into the limitations of next-token prediction in modeling human intelligence. Examine the critical distinction between autoregressive inference and teacher-forced training in language models. Discover why the popular criticism of error compounding during autoregressive inference may overlook a more fundamental issue: the potential failure of teacher-forcing to learn accurate next-token predictors for certain task classes. Investigate a general mechanism of teacher-forcing failure and analyze empirical evidence from a minimal planning task where both Transformer and Mamba architectures struggle. Consider the potential benefits of training models to predict multiple tokens in advance as a possible solution. Gain insights that can inform future debates and inspire research beyond the current next-token prediction paradigm in artificial intelligence.

Syllabus

The Pitfalls of Next-token Prediction


Taught by

Simons Institute

Related Courses

Launching into Machine Learning 日本語版
Google Cloud via Coursera
Launching into Machine Learning auf Deutsch
Google Cloud via Coursera
Launching into Machine Learning en Français
Google Cloud via Coursera
Launching into Machine Learning en Español
Google Cloud via Coursera
Основы машинного обучения
Higher School of Economics via Coursera