YoVDO

XGBoost Part 1 - Regression

Offered By: StatQuest with Josh Starmer via YouTube

Tags

XGBoost Courses Regularization Courses

Course Description

Overview

Dive into the first part of a four-part video series on XGBoost, focusing on its application to regression problems. Learn about the unique regression trees used in XGBoost, including initial predictions, tree building, similarity score calculations, gain evaluation for thresholds, tree pruning, regularization, output value calculations, and making predictions. Explore these concepts through clear explanations and visual aids, building on prior knowledge of Gradient Boost for Regression and Regularization. Gain a comprehensive understanding of XGBoost's approach to regression, preparing you for more advanced topics in subsequent videos.

Syllabus

Awesome song and introduction
The initial prediction
Building an XGBoost Tree for regression
Calculating Similarity Scores
Calculating Gain to evaluate different thresholds
Pruning an XGBoost Tree
Building an XGBoost Tree with regularization
Calculating output values for an XGBoost Tree
Making predictions with XGBoost
Summary of concepts and main ideas
I say "66", but I meant to say "62.48". However, either way, the conclusion is the same.
In the original XGBoost documents they use the epsilon symbol to refer to the learning rate, but in the actual implementation, this is controlled via the "eta" parameter. So, I guess to be consistent with the original documentation, I made the same mistake! :


Taught by

StatQuest with Josh Starmer

Related Courses

Data Analysis and Visualization
Georgia Institute of Technology via Udacity
Improving Deep Neural Networks: Hyperparameter Tuning, Regularization and Optimization
DeepLearning.AI via Coursera
機器學習基石下 (Machine Learning Foundations)---Algorithmic Foundations
National Taiwan University via Coursera
Data Science: Machine Learning
Harvard University via edX
Art and Science of Machine Learning auf Deutsch
Google Cloud via Coursera