NFNets - High-Performance Large-Scale Image Recognition Without Normalization
Offered By: Yannic Kilcher via YouTube
Course Description
Overview
Explore a comprehensive video analysis of the research paper "NFNets: High-Performance Large-Scale Image Recognition Without Normalization" from Google DeepMind. Delve into the innovative approach of Normalizer-Free Networks, which achieve state-of-the-art classification accuracy on ImageNet without using batch normalization. Learn about the advantages and disadvantages of BatchNorm, and discover how adaptive gradient clipping (AGC) and architectural improvements enable NFNets to outperform traditional models. Gain insights into the benefits of this new technique, including faster training, improved accuracy, and enhanced transfer learning performance. Follow along as the video breaks down the paper's key contributions, compares NFNets to EfficientNet, and discusses the implications for future deep learning research.
Syllabus
- Intro & Overview
- What's the problem with BatchNorm?
- Paper contribution Overview
- Beneficial properties of BatchNorm
- Previous work: NF-ResNets
- Adaptive Gradient Clipping
- AGC and large batch size
- AGC induces implicit dependence between training samples
- Are BatchNorm's problems solved?
- Network architecture improvements
- Comparison to EfficientNet
- Conclusion & Comments
Taught by
Yannic Kilcher
Related Courses
JAX Crash Course - Accelerating Machine Learning CodeAssemblyAI via YouTube Coding a Neural Network from Scratch in Pure JAX - Machine Learning with JAX - Tutorial 3
Aleksa Gordić - The AI Epiphany via YouTube Diffrax - Numerical Differential Equation Solvers in JAX
Fields Institute via YouTube JAX- Accelerated Machine Learning Research via Composable Function Transformations in Python
Fields Institute via YouTube Getting Started with Automatic Differentiation
PyCon US via YouTube