On the Connection Between Neural Networks and Kernels: A Modern Perspective - Simon Du
Offered By: Institute for Advanced Study via YouTube
Course Description
Overview
Explore a comprehensive lecture on the modern perspective of the connection between neural networks and kernels. Delve into fundamental questions, empirical observations on training loss and generalization, and over-parameterization in neural networks. Examine trajectory-based analysis, kernel matrices, and the main theory behind zero training error. Investigate empirical results on generalization, convolutional neural tangent kernels, and their application to CIFAR-10. Understand global and local average pooling, and explore experiments on UCI datasets and few-shot learning. Discover graph neural tangent kernels for graph classification, and gain insights into the latest research and references in this field.
Syllabus
Intro
Two Fundamental Questions
Empirical Observations on Training Loss
Over-parameterization
Empirical Observations on Generalization
Example: Two-layer NN
Trajectory-based Analysis
The Trajectory of Predictions (Cont'd)
Kernel Matrix at the Beginning
Kernel Matrix During Training
Main Theory
Zero Training Error
Empirical Results on Generalization
Convolutional Neural Tangent Kernel
CNTK on CIFAR 10
Understanding Global Average Pooling
Local Average Pooling
UCI Experiment Setup
UCI Results
Few-shot Learning Setup
Few-shot Learning Results
Graph NTK for Graph Classification
Summary
References
Taught by
Institute for Advanced Study
Related Courses
機器學習技法 (Machine Learning Techniques)National Taiwan University via Coursera Utilisez des modèles supervisés non linéaires
CentraleSupélec via OpenClassrooms Statistical Machine Learning
Eberhard Karls University of Tübingen via YouTube Interplay of Linear Algebra, Machine Learning, and HPC - JuliaCon 2021 Keynote
The Julia Programming Language via YouTube Interpolation and Learning With Scale Dependent Kernels
MITCBMM via YouTube