YoVDO

Stochastic Depth for Neural Networks - Implementation and Analysis

Offered By: Yacine Mahdid via YouTube

Tags

Neural Networks Courses Deep Learning Courses PyTorch Courses Network Architecture Courses Model Optimization Courses Regularization Courses

Course Description

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the powerful regularization method of stochastic depth for residual neural networks in this 27-minute tutorial. Learn about the methodology and its Pytorch implementation, covering topics such as architecture changes, differences between dropout and drop path, speedup and performance benefits, and analytical experiment results. Dive into a code walkthrough to gain practical insights, and understand how stochastic depth enables training shorter networks while using deep networks at test time. Discover how this approach complements residual networks, substantially reducing training time and significantly improving test error across various datasets.

Syllabus

- Introduction:
- Background and Context:
- Questions:
- Architecture Changes:
- Why use the resnet architecture for stochastic depth?:
- Difference between Drop out and Drop path:
- Speedup and performance:
- Stochastic Depth example:
- How do they manage speedup and better performance with stochastic depth?:
- Data sets:
- Main Results:
- Analytical Experiment Result:
- Code Walkthrough:
- Conclusion:


Taught by

Yacine Mahdid

Related Courses

An Introduction to Computer Networks
Stanford University via Independent
Introduction aux réseaux cellulaires
Institut Mines-Télécom via Independent
Introduction aux réseaux mobiles
Institut Mines-Télécom via France Université Numerique
Comprendre la 4G
Institut Mines-Télécom via France Université Numerique
4G Network Essentials
Institut Mines-Télécom via edX