AdaBoost, Clearly Explained
Offered By: StatQuest with Josh Starmer via YouTube
Course Description
Overview
Dive into a comprehensive video tutorial that demystifies AdaBoost, a powerful machine learning algorithm. Learn how this method builds upon decision trees and random forests to create a robust ensemble model. Explore the three main ideas behind AdaBoost, including building stumps with the GINI index, determining the "Amount of Say" for each stump, and updating sample weights. Follow along as the tutorial guides you through the process of normalizing weights, creating subsequent stumps, and using the ensemble to make classifications. Gain a clear understanding of AdaBoost's inner workings through step-by-step explanations, visual aids, and a thorough review of key concepts.
Syllabus
Awesome song and introduction
The three main ideas behind AdaBoost
Review of the three main ideas
Building a stump with the GINI index
Determining the Amount of Say for a stump
Updating sample weights
Normalizing the sample weights
Using the normalized weights to make the second stump
Using stumps to make classifications
Review of the three main ideas behind AdaBoost
. The Amount of Say for Chest Pain = 1/2*log1-3/8/3/8 = 1/2*log5/8/3/8 = 1/2*log5/3 = 0.25, not 0.42.
Taught by
StatQuest with Josh Starmer
Related Courses
Advanced AI Techniques for the Supply ChainLearnQuest via Coursera Build Decision Trees, SVMs, and Artificial Neural Networks
CertNexus via Coursera Ensemble Methods in Machine Learning
Codecademy Machine Learning: Random Forests & Decision Trees
Codecademy Advanced Methods in Machine Learning Applications
Johns Hopkins University via Coursera