YoVDO

Decision Transformer - Reinforcement Learning via Sequence Modeling

Offered By: Yannic Kilcher via YouTube

Tags

Offline Reinforcement Learning Courses Reinforcement Learning Courses Sequence Modeling Courses Transformer Architecture Courses

Course Description

Overview

Explore a comprehensive video explanation of the research paper "Decision Transformer: Reinforcement Learning via Sequence Modeling." Delve into the innovative approach of framing offline reinforcement learning as a sequence modeling problem, leveraging the power of Transformer architectures. Learn about the Decision Transformer model, which generates optimal actions by conditioning on desired returns, past states, and actions. Discover how this method compares to traditional value function and policy gradient approaches in reinforcement learning. Examine key concepts such as offline reinforcement learning, temporal difference learning, reward-to-go, and the context length problem. Analyze experimental results on various benchmarks and gain insights into the potential implications of this research for the field of reinforcement learning.

Syllabus

- Intro & Overview
- Offline Reinforcement Learning
- Transformers in RL
- Value Functions and Temporal Difference Learning
- Sequence Modeling and Reward-to-go
- Why this is ideal for offline RL
- The context length problem
- Toy example: Shortest path from random walks
- Discount factors
- Experimental Results
- Do you need to know the best possible reward?
- Key-to-door toy experiment
- Comments & Conclusion


Taught by

Yannic Kilcher

Related Courses

Can Wikipedia Help Offline Reinforcement Learning - Author Interview
Yannic Kilcher via YouTube
Can Wikipedia Help Offline Reinforcement Learning? - Paper Explained
Yannic Kilcher via YouTube
CAP6412 - Final Project Presentations - Lecture 27
University of Central Florida via YouTube
Offline Reinforcement Learning and Model-Based Optimization
Simons Institute via YouTube
Reinforcement Learning
Simons Institute via YouTube