YoVDO

Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift

Offered By: Yannic Kilcher via YouTube

Tags

Batch Normalization Courses Machine Learning Courses Neural Networks Courses Regularization Courses

Course Description

Overview

Explore a comprehensive video analysis of the groundbreaking paper "Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift." Delve into the concept of internal covariate shift and its impact on deep neural network training. Learn how batch normalization addresses this issue by normalizing layer inputs, allowing for higher learning rates and less careful parameter initialization. Discover how this technique acts as a regularizer, potentially eliminating the need for dropout. Examine the impressive results achieved when applying batch normalization to state-of-the-art image classification models, including significant improvements in training speed and accuracy. Gain insights into the paper's methodology, implementation details, and its impact on the field of deep learning.

Syllabus

Introduction
What is Batch Normalization
Training
Back Propagation
Results


Taught by

Yannic Kilcher

Related Courses

Introduction to Artificial Intelligence
Stanford University via Udacity
Natural Language Processing
Columbia University via Coursera
Probabilistic Graphical Models 1: Representation
Stanford University via Coursera
Computer Vision: The Fundamentals
University of California, Berkeley via Coursera
Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent