YoVDO

Two Stories in Mechanistic Interpretation of Natural and Artificial Neural Computation

Offered By: Institute for Pure & Applied Mathematics (IPAM) via YouTube

Tags

Transformers Courses Artificial Intelligence Courses Linear Regression Courses Generalization Courses In-context Learning Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a 55-minute conference talk by Cengiz Pehlevan from Harvard University, presented at IPAM's Theory and Practice of Deep Learning Workshop. Delve into two stories of mechanistic interpretation in natural and artificial neural computation. Examine the remarkable ability of Transformers to perform in-context learning (ICL) without explicit prior training. Investigate an exactly solvable model of ICL for linear regression tasks using linear attention, uncovering sharp asymptotics for the learning curve in a scaling regime with infinite token dimension. Discover a double-descent learning curve and a phase transition between low and high task diversity regimes, revealing insights into memorization versus genuine in-context learning and generalization. Validate theoretical findings through experiments with both linear attention and full nonlinear Transformer architectures.

Syllabus

Cengiz Pehlevan - 2 stories in mechanistic interpretation of natural & artificial neural computation


Taught by

Institute for Pure & Applied Mathematics (IPAM)

Related Courses

CMU Advanced NLP: How to Use Pre-Trained Models
Graham Neubig via YouTube
Stanford Seminar 2022 - Transformer Circuits, Induction Heads, In-Context Learning
Stanford University via YouTube
Pretraining Task Diversity and the Emergence of Non-Bayesian In-Context Learning for Regression
Simons Institute via YouTube
In-Context Learning: A Case Study of Simple Function Classes
Simons Institute via YouTube
AI Mastery: Ultimate Crash Course in Prompt Engineering for Large Language Models
Data Science Dojo via YouTube