YoVDO

EfficientNetV2 - Smaller Models and Faster Training - Paper Explained

Offered By: Aleksa Gordić - The AI Epiphany via YouTube

Tags

Neural Networks Courses Model Optimization Courses Neural Architecture Search Courses

Course Description

Overview

Explore a comprehensive video explanation of the EfficientNetV2 paper, which introduces smaller models and faster training techniques for image classification. Learn about progressive training, the Fused-MBConv layer, and a novel reward function for Neural Architecture Search (NAS). Dive deep into the paper's key concepts, including a high-level overview, NAS review, novel reward function, progressive training, stochastic depth regularization, and results. Gain insights into how EfficientNetV2 achieves better performance on ImageNet top-1 accuracy compared to recent models like NFNets and Vision Transformers.

Syllabus

High-level overview
NAS review
Deep dive
Novel reward
Progressive training
Stochastic depth regularization
Results


Taught by

Aleksa Gordić - The AI Epiphany

Related Courses

Machine Learning Modeling Pipelines in Production
DeepLearning.AI via Coursera
MLOps for Scaling TinyML
Harvard University via edX
Parameter Prediction for Unseen Deep Architectures - With First Author Boris Knyazev
Yannic Kilcher via YouTube
SpineNet - Learning Scale-Permuted Backbone for Recognition and Localization
Yannic Kilcher via YouTube
Synthetic Petri Dish - A Novel Surrogate Model for Rapid Architecture Search
Yannic Kilcher via YouTube