What Are the Statistical Limits of Offline Reinforcement Learning With Function Approximation?
Offered By: Simons Institute via YouTube
Course Description
Overview
Explore the statistical boundaries of offline reinforcement learning with function approximation in this 55-minute lecture by Sham Kakade from the University of Washington and Microsoft Research. Delve into key concepts including realizability, sequential decision making, coverage limits, and policy evaluation. Examine upper and lower bounds, practical considerations, and experimental results. Gain insights into the mathematics of online decision making and the interplay between models and features in reinforcement learning.
Syllabus
Intro
What is offline reinforcement learning
Intuition
Realizability
Sequential Decision Making
Standard Approach
Coverage
Limits
Policy Evaluation
Setting
Feature Mapping
Upper Limits
Lower Limits
Observations
Upper Bounds
Inequality
Simulation
Summary
Sufficient Conditions
Possible Results
Intuition and Construction
Practical Considerations
Follow Up
Experiments
Other Experiments
Model vs Feature
Taught by
Simons Institute
Related Courses
Introduction to Operations ManagementWharton School of the University of Pennsylvania via Coursera Computational Molecular Evolution
Technical University of Denmark (DTU) via Coursera Structural Equation Model and its Applications | 结构方程模型及其应用 (普通话)
The Chinese University of Hong Kong via Coursera Fundamentals of Clinical Trials
Harvard University via edX Curso Práctico de Bioestadística con R
Universidad San Pablo CEU via Miríadax