Theoretical and Practical Insights from Linear Transformers
Offered By: Simons Institute via YouTube
Course Description
Overview
Explore theoretical and practical insights into Linear Transformers in this 34-minute lecture by Xiang Cheng from the Massachusetts Institute of Technology. Delve into recent research highlighting Linear Transformers as proxies for understanding full-fledged Transformer models. Examine theoretical proofs demonstrating how Linear Transformers learn linear regression tasks in-context through gradient-based optimization during forward passes. Gain insights into the mechanisms behind Transformers' in-context learning capabilities. Discover intriguing empirical observations suggesting that the optimization landscape of Linear Transformers may serve as a valuable approximation for understanding the optimization of real Transformers. Enhance your knowledge of optimization and algorithm design in the context of transformer models.
Syllabus
Theoretical and Practical Insights from Linear Transformers
Taught by
Simons Institute
Related Courses
Natural Language ProcessingColumbia University via Coursera Intro to Algorithms
Udacity Conception et mise en œuvre d'algorithmes.
École Polytechnique via Coursera Paradigms of Computer Programming
Université catholique de Louvain via edX Data Structures and Algorithm Design Part I | 数据结构与算法设计(上)
Tsinghua University via edX