YoVDO

Tensor Programs - Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer

Offered By: Aleksa Gordić - The AI Epiphany via YouTube

Tags

Neural Networks Courses Zero-shot learning (ZSL) Courses Hyperparameter Optimization Courses

Course Description

Overview

Dive into an in-depth exploration of the "Tensor Programs V: Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer (μTransfer)" paper in this comprehensive video lecture. Learn about the groundbreaking approach that makes optimal hyperparameters stable with respect to width scaling. Explore previous work on tensor programs, revisit the neural tangent kernel (NTK) concept, and understand the abc parametrization. Discover how learning occurs in NTK and its connections to the Central Limit Theorem. Gain practical insights into Maximal Update Parametrization and its relationship to the DeepNet paper. Analyze the results that suggest width might be the key factor in neural network performance. Enhance your understanding of advanced machine learning concepts and cutting-edge research in neural network optimization.

Syllabus

uTransfer introduced
Previous work tensor programs IV
NTK - neural tangent kernel recap
abc parametrization
How does learning happen in NTK?
Connections to Central Limit Theorem
Maximal Update Parametrization in Practice
DeepNet paper connection
Results width is all you need?


Taught by

Aleksa Gordić - The AI Epiphany

Related Courses

Improving Deep Neural Networks: Hyperparameter Tuning, Regularization and Optimization
DeepLearning.AI via Coursera
How to Win a Data Science Competition: Learn from Top Kagglers
Higher School of Economics via Coursera
Predictive Modeling and Machine Learning with MATLAB
MathWorks via Coursera
Machine Learning Rapid Prototyping with IBM Watson Studio
IBM via Coursera
Hyperparameter Tuning with Neural Network Intelligence
Coursera Project Network via Coursera