YoVDO

Axial-DeepLab - Stand-Alone Axial-Attention for Panoptic Segmentation

Offered By: Yannic Kilcher via YouTube

Tags

Self-Attention Mechanisms Courses Artificial Intelligence Courses Machine Learning Courses Image Processing Courses ResNet Courses

Course Description

Overview

Explore a comprehensive video explanation of the research paper "Axial-DeepLab: Stand-Alone Axial-Attention for Panoptic Segmentation". Delve into the innovative approach of combining learned Positional Embeddings with Axial Attention to create a fully attentional model for image processing. Learn how this new model competes with Convolutional Neural Networks in image classification and achieves state-of-the-art results in various image segmentation tasks. Follow the detailed breakdown of concepts, from the transition from convolution to self-attention in images, to the implementation of axial attention and its application in replacing convolutions in ResNet. Gain insights into the experimental results and practical examples demonstrating the model's effectiveness across multiple large-scale datasets.

Syllabus

- Intro & Overview
- This Paper's Contributions
- From Convolution to Self-Attention for Images
- Learned Positional Embeddings
- Propagating Positional Embeddings through Layers
- Traditional vs Position-Augmented Attention
- Axial Attention
- Replacing Convolutions in ResNet
- Experimental Results & Examples


Taught by

Yannic Kilcher

Related Courses

Linformer - Self-Attention with Linear Complexity
Yannic Kilcher via YouTube
Synthesizer - Rethinking Self-Attention in Transformer Models
Yannic Kilcher via YouTube
The Narrated Transformer Language Model
Jay Alammar via YouTube
Learning the Structure of EHR with Graph Convolutional Transformer - Edward Choi
Stanford University via YouTube
Knowledge Is Embedded in Language Neural Networks but Can They Reason?
Simons Institute via YouTube