Bayesian and Contextual Bandits
Offered By: Pascal Poupart via YouTube
Course Description
Overview
Explore the fundamentals of Bayesian and contextual bandits in this comprehensive lecture. Delve into multi-armed bandits, Bayesian learning, and distributional information before examining a coin example to understand Bernoulli variables and Beta distribution. Compare approaches and learn about belief updates before diving into the Thompson Sampling Algorithm for Bernoulli rewards. Conclude by investigating contextual bandits, approximate reward functions, and predictive posteriors, gaining a solid foundation in these crucial machine learning concepts.
Syllabus
Intro
Multi-Armed Bandits
Bayesian Learning
Distributional Information
Coin Example
Bernoulli Variables
Beta distribution
Comparison
Belief Update
Thompson Sampling Algorithm Bernoulli Rewards
Contextual Bandits
Approximate Reward Function
Predictive Posterior
Taught by
Pascal Poupart
Related Courses
Computational NeuroscienceUniversity of Washington via Coursera Reinforcement Learning
Brown University via Udacity Reinforcement Learning
Indian Institute of Technology Madras via Swayam FA17: Machine Learning
Georgia Institute of Technology via edX Introduction to Reinforcement Learning
Higher School of Economics via Coursera