Computer Vision Architecture Evolution: ConvNets to Transformers - Lecture 21
Offered By: University of Central Florida via YouTube
Course Description
Overview
Explore the evolution of vision architectures and modernization of convolutional neural networks in this 33-minute lecture from the University of Central Florida. Delve into the hierarchy of SWIN versus CNNs, macro design changes in ResNet, and improvements like inverted bottlenecks and larger kernel sizes. Examine micro designs, including activation function replacements and normalization layer adjustments. Learn about the final ConvNext block, network evaluation techniques, and compare machine performance across different architectures.
Syllabus
Introduction
Evolution of Vision Architectures
Hierarchy of SWIN vs. CNNs
Modernizing ConvNets
Modernizing ResNet
Macro Design Changes
Changing stage compute ratio
Changing stem to "Patch-ify"
Depthwise Conv. vs Self-Attention
Improvements
Inverted Bottleneck
Larger Kernel Sizes
Micro Designs (mD)
Replace RELU with GELU
Fewer Activation functions
Fewer Normalization Layers
Substituting BN with LN
Visualization
mD4- Improvement
Separate Downsampling Layer
Final ConvNext block
Networks for Evaluation
Training Settings
Machine Performance Comparison
Taught by
UCF CRCV
Tags
Related Courses
Creating Multi Task Models With KerasCoursera Project Network via Coursera Deep Learning: Advanced Computer Vision (GANs, SSD, +More!)
Udemy 【Hands Onで学ぶ】PyTorchによる深層学習入門
Udemy Axial-DeepLab - Stand-Alone Axial-Attention for Panoptic Segmentation
Yannic Kilcher via YouTube Deep Residual Learning for Image Recognition - Paper Explained
Yannic Kilcher via YouTube