YoVDO

Parameter Sharing - Recurrent and Convolutional Nets

Offered By: Alfredo Canziani via YouTube

Tags

Neural Networks Courses Deep Learning Courses Long short-term memory (LSTM) Courses Attention Mechanisms Courses Hypernetworks Courses

Course Description

Overview

Explore parameter sharing in recurrent and convolutional neural networks in this comprehensive 2-hour lecture by Yann LeCun. Delve into hypernetworks, shared weights, and gradient addition in parameter sharing. Examine recurrent nets, including unrolling in time, vanishing and exploding gradients, and RNN tricks. Investigate memory concepts, LSTM networks, and attention mechanisms for sequence-to-sequence mapping. Study convolutional nets, including motif detection, convolution definitions, backpropagation, and architecture. Learn about vintage ConvNets, brain image interpretation, and the Hubel & Wiesel model of the visual cortex. Gain insights into ConvNet invariance and equivariance, training time, iteration cycles, and historical remarks in deep learning.

Syllabus

– Welcome to class
– Hypernetworks
– Shared weights
– Parameter sharing ⇒ adding the gradients
– Max and sum reductions
– Recurrent nets
– Unrolling in time
– Vanishing and exploding gradients
– Math on the whiteboard
– RNN tricks
– RNN for differential equations
– GRU
– What is a memory
– LSTM – Long Short-Term Memory net
– Multilayer LSTM
– Attention for sequence to sequence mapping
– Convolutional nets
– Detecting motifs in images
– Convolution definitions
– Backprop through convolutions
– Stride and skip: subsampling and convolution “à trous”
– Convolutional net architecture
– Multiple convolutions
– Vintage ConvNets
– How does the brain interpret images?
– Hubel & Wiesel's model of the visual cortex
– Invariance and equivariance of ConvNets
– In the next episode…
– Training time, iteration cycle, and historical remarks


Taught by

Alfredo Canziani

Tags

Related Courses

SIREN - Implicit Neural Representations with Periodic Activation Functions
Yannic Kilcher via YouTube
Textual Inversion and Hypernetworks - Stable Diffusion 2
Nerdy Rodent via YouTube
Stable Diffusion Style Technique Comparison - Hypernetwork vs. Textual Inversion
kasukanra via YouTube
Emergent Hypernetworks in Weakly Coupled Oscillators - IPAM at UCLA
Institute for Pure & Applied Mathematics (IPAM) via YouTube
Improving Pareto Front Learning via Multi-Head HyperNetwork
VinAI via YouTube