Beyond Empirical Risk Minimization - The Lessons of Deep Learning
Offered By: MITCBMM via YouTube
Course Description
Overview
Syllabus
Intro
The ERM/SRM theory of learning
Unifom laws of large numbers
Capacity control
U-shaped generalization curve
Does interpolation overfit?
Interpolation does not averfit even for very noisy data
why bounds fail
Interpolation is best practice for deep learning
Historical recognition
The key lesson
Generalization theory for interpolation?
A way forward?
Interpolated k-NN schemes
Interpolation and adversarial examples
Double descent risk curve
More parameters are better: an example
Random Fourier networks
what is the mechanism?
Double Descent in Randon Feature settings
Smoothness by averaging
Framework for modern ML
The landscape of generalization
Optimization: classical
Modern Optimization
From classical statistics to modern ML
The nature of inductive bias
Memorization and interpolation
Interpolation in deep auto-encoders
Neural networks as models for associative memory
Why are attractors surprising?
Memorizing sequences
Taught by
MITCBMM
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Natural Language Processing
Columbia University via Coursera Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent