Supermasks in Superposition - Paper Explained
Offered By: Yannic Kilcher via YouTube
Course Description
Overview
Explore an in-depth analysis of the research paper "Supermasks in Superposition" in this comprehensive video lecture. Delve into the concept of supermasks, binary masks of randomly initialized neural networks that perform well on specific tasks, and their application in lifelong learning. Learn how the system can automatically derive task IDs at inference time and distinguish up to 2500 tasks. Follow along as the lecture covers key topics including catastrophic forgetting, mask superpositions, binary maximum entropy search, and encoding masks in Hopfield networks. Gain insights into the paper's methodology, experiments, and conclusions, as well as potential applications and extensions of this innovative approach to sequential learning in neural networks.
Syllabus
- Intro & Overview
- Catastrophic Forgetting
- Supermasks
- Lifelong Learning using Supermasks
- Inference Time Task Discrimination by Entropy
- Mask Superpositions
- Proof-of-Concept, Task Given at Inference
- Binary Maximum Entropy Search
- Task Not Given at Inference
- Task Not Given at Training
- Ablations
- Superfluous Neurons
- Task Selection by Detecting Outliers
- Encoding Masks in Hopfield Networks
- Conclusion
Taught by
Yannic Kilcher
Related Courses
Active Dendrites Avoid Catastrophic Forgetting - Interview With the AuthorsYannic Kilcher via YouTube Avoiding Catastrophe - Active Dendrites Enable Multi-Task Learning in Dynamic Environments
Yannic Kilcher via YouTube What Kind of AI Can Help Manufacturing Adapt to a Pandemic
Open Data Science via YouTube Rethinking Architecture Design for Data Heterogeneity in FL - Liangqiong Qu
Stanford University via YouTube Continual Learning and Catastrophic Forgetting
Paul Hand via YouTube