YoVDO

What Are the Statistical Limits of Offline Reinforcement Learning With Function Approximation?

Offered By: Simons Institute via YouTube

Tags

Offline Reinforcement Learning Courses Statistical Analysis Courses Function Approximation Courses Sequential Decision Making Courses

Course Description

Overview

Explore the statistical boundaries of offline reinforcement learning with function approximation in this 55-minute lecture by Sham Kakade from the University of Washington and Microsoft Research. Delve into key concepts including realizability, sequential decision making, coverage limits, and policy evaluation. Examine upper and lower bounds, practical considerations, and experimental results. Gain insights into the mathematics of online decision making and the interplay between models and features in reinforcement learning.

Syllabus

Intro
What is offline reinforcement learning
Intuition
Realizability
Sequential Decision Making
Standard Approach
Coverage
Limits
Policy Evaluation
Setting
Feature Mapping
Upper Limits
Lower Limits
Observations
Upper Bounds
Inequality
Simulation
Summary
Sufficient Conditions
Possible Results
Intuition and Construction
Practical Considerations
Follow Up
Experiments
Other Experiments
Model vs Feature


Taught by

Simons Institute

Related Courses

Introduction to Operations Management
Wharton School of the University of Pennsylvania via Coursera
Computational Molecular Evolution
Technical University of Denmark (DTU) via Coursera
Structural Equation Model and its Applications | 结构方程模型及其应用 (普通话)
The Chinese University of Hong Kong via Coursera
Fundamentals of Clinical Trials
Harvard University via edX
Curso Práctico de Bioestadística con R
Universidad San Pablo CEU via Miríadax