YoVDO

Mamba: Linear-Time Sequence Modeling with Selective State Spaces

Offered By: Yannic Kilcher via YouTube

Tags

Machine Learning Courses Deep Learning Courses Neural Networks Courses Algorithm Design Courses Transformers Courses Sequence Modeling Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a comprehensive video lecture on the Mamba architecture, a novel approach to linear-time sequence modeling using selective state spaces. Delve into the comparison between Transformers, RNNs, and S4 models before examining state space models and their selective variants. Analyze the Mamba architecture in detail, including its SSM layer and forward propagation techniques. Discover how the model utilizes GPU memory hierarchy and achieves efficient computation through prefix sums and parallel scans. Review experimental results, gain insights from the presenter's comments, and conclude with a brief examination of the underlying code. Enhance your understanding of this cutting-edge approach to sequence modeling that outperforms Transformers in various modalities while offering faster inference and linear scaling in sequence length.

Syllabus

- Introduction
- Transformers vs RNNs vs S4
- What are state space models?
- Selective State Space Models
- The Mamba architecture
- The SSM layer and forward propagation
- Utilizing GPU memory hierarchy
- Efficient computation via prefix sums / parallel scans
- Experimental results and comments
- A brief look at the code


Taught by

Yannic Kilcher

Related Courses

Natural Language Processing
Columbia University via Coursera
Intro to Algorithms
Udacity
Conception et mise en œuvre d'algorithmes.
École Polytechnique via Coursera
Paradigms of Computer Programming
Université catholique de Louvain via edX
Data Structures and Algorithm Design Part I | 数据结构与算法设计(上)
Tsinghua University via edX