Approximate Cross Validation for Large Data and High Dimensions - Tamara Broderick, MIT
Offered By: Alan Turing Institute via YouTube
Course Description
Overview
Explore a 35-minute lecture on approximate cross-validation techniques for large datasets and high-dimensional problems. Delve into the challenges of assessing error and variability in statistical and machine learning algorithms when dealing with modern, large-scale data. Learn about the infinitesimal jackknife (IJ) method, a linear approximation technique that can significantly speed up cross-validation and bootstrap processes. Examine finite-sample error bounds for the IJ and discover how dimensionality reduction can be applied to improve its performance in high-dimensional scenarios, particularly for leave-one-out cross-validation (LOOCV) with L1 regularization in generalized linear models. Gain insights into the theoretical foundations and practical applications of these techniques through simulated and real-data experiments, and understand how they contribute to the growing intersection of statistics and computer science in the field of machine learning.
Syllabus
Introduction
Crossvalidation setup
Firstorder Taylor expansion
Package
Theory
Assumptions
Experiments
Results
Conclusions
Taught by
Alan Turing Institute
Related Courses
Продвинутые методы машинного обученияHigher School of Economics via Coursera Natural Language Processing with Classification and Vector Spaces
DeepLearning.AI via Coursera Machine Learning - Dimensionality Reduction
IBM via Cognitive Class Machine Learning with Python
IBM via Cognitive Class Predicting Extreme Climate Behavior with Machine Learning
University of Colorado Boulder via Coursera