Provably Efficient Reinforcement Learning with Linear Function Approximation - Chi Jin
Offered By: Institute for Advanced Study via YouTube
Course Description
Overview
Explore provably efficient reinforcement learning with linear function approximation in this 28-minute lecture from the Workshop on Theory of Deep Learning. Delve into sequential decision making, sample efficiency, and value-based algorithms as Chi Jin, a Member of the School of Mathematics at the Institute for Advanced Study, presents cutting-edge research. Examine exploration techniques, including multi-armed bandits and Upper Confidence Bound (UCB), before moving beyond tabular settings to linear function approximation. Investigate linear MDPs and related work in this comprehensive overview of reinforcement learning theory and applications.
Syllabus
Intro
Sequential Decision Making
Reinforcement Learning
Sample Efficiency
Value-based Algorithms
Exploration
Multi-armed Bandits
Upper Confidence Bound (UCB)
Q-learning with UCB
Beyond Tabular Setting
Linear Function Approximation
A Natural Algorithm
Linear MDP
Related Work
Taught by
Institute for Advanced Study
Related Courses
Toward Generalizable Embodied AI for Machine AutonomyBolei Zhou via YouTube What Are the Statistical Limits of Offline Reinforcement Learning With Function Approximation?
Simons Institute via YouTube Better Learning from the Past - Counterfactual - Batch RL
Simons Institute via YouTube Off-Policy Policy Optimization
Simons Institute via YouTube Divide-and-Conquer Monte Carlo Tree Search for Goal-Directed Planning - Paper Explained
Yannic Kilcher via YouTube