YoVDO

Provably Efficient Reinforcement Learning with Linear Function Approximation - Chi Jin

Offered By: Institute for Advanced Study via YouTube

Tags

Reinforcement Learning Courses Q-learning Courses Multi-Armed Bandits Courses Sequential Decision Making Courses

Course Description

Overview

Explore provably efficient reinforcement learning with linear function approximation in this 28-minute lecture from the Workshop on Theory of Deep Learning. Delve into sequential decision making, sample efficiency, and value-based algorithms as Chi Jin, a Member of the School of Mathematics at the Institute for Advanced Study, presents cutting-edge research. Examine exploration techniques, including multi-armed bandits and Upper Confidence Bound (UCB), before moving beyond tabular settings to linear function approximation. Investigate linear MDPs and related work in this comprehensive overview of reinforcement learning theory and applications.

Syllabus

Intro
Sequential Decision Making
Reinforcement Learning
Sample Efficiency
Value-based Algorithms
Exploration
Multi-armed Bandits
Upper Confidence Bound (UCB)
Q-learning with UCB
Beyond Tabular Setting
Linear Function Approximation
A Natural Algorithm
Linear MDP
Related Work


Taught by

Institute for Advanced Study

Related Courses

Toward Generalizable Embodied AI for Machine Autonomy
Bolei Zhou via YouTube
What Are the Statistical Limits of Offline Reinforcement Learning With Function Approximation?
Simons Institute via YouTube
Better Learning from the Past - Counterfactual - Batch RL
Simons Institute via YouTube
Off-Policy Policy Optimization
Simons Institute via YouTube
Divide-and-Conquer Monte Carlo Tree Search for Goal-Directed Planning - Paper Explained
Yannic Kilcher via YouTube