Sub-Linear Algorithms Meets Large Language Models
Offered By: Simons Institute via YouTube
Course Description
Overview
Explore the intersection of sub-linear algorithms and large language models in this 41-minute talk by Anshumali Shrivastava from Rice University. Delve into the challenges faced by large language models (LLMs) as their compute, memory, and energy requirements reach trillions of levels per input. Examine the limitations of Large Context Attention and KV Cache blowup, and discover why breaking linear resource barriers is crucial for advancing LLMs. Learn about emerging ideas and successful trends in applying sub-linear algorithms to future LLMs, emphasizing their necessity rather than optionality in pushing the boundaries of language model capabilities.
Syllabus
Sub-Linear Algorithms meets Large Language Models
Taught by
Simons Institute
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Natural Language Processing
Columbia University via Coursera Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent