YoVDO

Approximate Cross Validation for Large Data and High Dimensions - Tamara Broderick, MIT

Offered By: Alan Turing Institute via YouTube

Tags

Dimensionality Reduction Courses Linear Approximation Courses

Course Description

Overview

Explore a 35-minute lecture on approximate cross-validation techniques for large datasets and high-dimensional problems. Delve into the challenges of assessing error and variability in statistical and machine learning algorithms when dealing with modern, large-scale data. Learn about the infinitesimal jackknife (IJ) method, a linear approximation technique that can significantly speed up cross-validation and bootstrap processes. Examine finite-sample error bounds for the IJ and discover how dimensionality reduction can be applied to improve its performance in high-dimensional scenarios, particularly for leave-one-out cross-validation (LOOCV) with L1 regularization in generalized linear models. Gain insights into the theoretical foundations and practical applications of these techniques through simulated and real-data experiments, and understand how they contribute to the growing intersection of statistics and computer science in the field of machine learning.

Syllabus

Introduction
Crossvalidation setup
Firstorder Taylor expansion
Package
Theory
Assumptions
Experiments
Results
Conclusions


Taught by

Alan Turing Institute

Related Courses

Продвинутые методы машинного обучения
Higher School of Economics via Coursera
Natural Language Processing with Classification and Vector Spaces
DeepLearning.AI via Coursera
Machine Learning - Dimensionality Reduction
IBM via Cognitive Class
Machine Learning with Python
IBM via Cognitive Class
Predicting Extreme Climate Behavior with Machine Learning
University of Colorado Boulder via Coursera