Networks that Adapt to Intrinsic Dimensionality Beyond the Domain
Offered By: Inside Livermore Lab via YouTube
Course Description
Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the intricacies of deep learning networks and their ability to adapt to intrinsic dimensionality in this seminar by Alexander Cloninger from UC San Diego. Delve into the central question of network size requirements for function approximation and how data dimensionality impacts learning. Examine ReLU networks' approximation capabilities for functions with dimensionality-reducing feature maps, focusing on projections onto low-dimensional submanifolds and distances to low-dimensional sets. Discover how deep nets remain faithful to an intrinsic dimension governed by the function rather than domain complexity. Investigate connections to two-sample testing, manifold autoencoders, and data generation. Learn about Dr. Cloninger's research in geometric data analysis and applied harmonic analysis, exploring applications in imaging, medicine, and artificial intelligence.
Syllabus
Introduction
Speaker Introduction
Overview
Neural Networks
The Curse of Dimensionality
Theory
Main Question
Manifold Learning Community
Reach of a Manifold
Linear Regression
Approximation Theory
Classification
Excess Risk
Recent Work
Chart Auto Encoders
Neural Network Construction
Linear Encoders
Clustered Data
Questions
Conclusion
Hybrid Seminar
Taught by
Inside Livermore Lab
Related Courses
Продвинутые методы машинного обученияHigher School of Economics via Coursera Natural Language Processing with Classification and Vector Spaces
DeepLearning.AI via Coursera Machine Learning - Dimensionality Reduction
IBM via Cognitive Class Machine Learning with Python
IBM via Cognitive Class Predicting Extreme Climate Behavior with Machine Learning
University of Colorado Boulder via Coursera