Continuous-in-time Limit for Bandits
Offered By: USC Probability and Statistics Seminar via YouTube
Course Description
Overview
Explore the connection between Hamilton-Jacobi-Bellman equations and multi-armed bandit (MAB) problems in this 44-minute seminar talk from the USC Probability and Statistics Seminar series. Delve into the first work establishing this connection in a general setting, as presented by Yuhua Zhu from UCSD. Learn about an efficient algorithm for solving MAB problems based on this newly established link and discover its practical applications. Gain insights into the exploration-exploitation trade-off in sequential decision making under uncertainty, a key concept in MAB paradigms.
Syllabus
Yuhua Zhu: Continuous-in-time Limit for Bandits (UCSD)
Taught by
USC Probability and Statistics Seminar
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Statistics One
Princeton University via Coursera Intro to Statistics
Stanford University via Udacity Passion Driven Statistics
Wesleyan University via Coursera