YoVDO

Stochastic Bandits: Foundations and Current Perspectives

Offered By: Simons Institute via YouTube

Tags

Thompson Sampling Courses

Course Description

Overview

Explore the foundations and current perspectives of stochastic bandits in this comprehensive lecture by Shipra Agrawal from Columbia University. Delve into the fundamental model of sequential learning, where rewards from different actions are assumed to come identically and independently from fixed distributions. Gain insights into the main algorithms for stochastic bandits, including Upper Confidence Bound and Thompson Sampling. Discover how these algorithms can be adapted to incorporate various additional constraints. This talk, part of the Data-Driven Decision Processes Boot Camp at the Simons Institute, provides a thorough examination of this crucial topic in sequential learning and decision-making processes.

Syllabus

Stochastic Bandits: Foundations and Current Perspectives


Taught by

Simons Institute

Related Courses

Artificial Intelligence for Business + ChatGPT Prize [2024]
Udemy
Probability
Serrano.Academy via YouTube
Machine Learning
Serrano.Academy via YouTube
Deep Bayesian Bandits - Exploring in Online Personalized Recommendations
Launchpad via YouTube
Online Learning and Bandits - Part 2
Simons Institute via YouTube