YoVDO

CS885: Multi-Armed Bandits

Offered By: Pascal Poupart via YouTube

Tags

Reinforcement Learning Courses Heuristics Courses Decision-Making Algorithms Courses Multi-Armed Bandits Courses

Course Description

Overview

Explore the fascinating world of multi-armed bandits in this comprehensive 57-minute lecture by Pascal Poupart. Delve into key concepts such as exploration-exploitation trade-offs, stochastic bandits, and online optimization. Learn about the origins of bandits in gambling and their practical applications. Understand the simplified version of the problem, various heuristics, and the notion of regret. Discover the epsilon-greedy strategy and its implementation in single-state scenarios. Gain insights into different approaches and their effectiveness in real-world situations.

Syllabus

Multiarmed bandits
Exploration exploitation
Stochastic bandits
Bandits from gambling
Bandits in practice
Online optimization
Simplified version
The problem
Heuristics
Notion of regret
Epsilon greedy strategy
Single state
Epsilon greedy
Different approaches
In practice


Taught by

Pascal Poupart

Related Courses

Fundamentals of Reinforcement Learning
University of Alberta via Coursera
Arduino Step by Step More than 50 Hours Complete Course
Udemy
Tic Tac Toe Tutorial in Python
Tech with Tim via YouTube
Explaining Decision-Making Algorithms through UI - Strategies to Help Non-Expert Stakeholders
Association for Computing Machinery (ACM) via YouTube
Algorithmic Bias in AI: Implications and Challenges - A Discussion with Cristopher Moore and Melanie Moses
Santa Fe Institute via YouTube