YoVDO

Blockwise Parallel Decoding for Deep Autoregressive Models

Offered By: Yannic Kilcher via YouTube

Tags

Deep Learning Courses Image Processing Courses

Course Description

Overview

Explore a novel blockwise parallel decoding scheme for deep autoregressive sequence-to-sequence models in this informative video. Learn how this approach allows for substantial improvements in generation speed when applied to architectures that can process output sequences in parallel. Discover the empirical verification of this method through experiments using state-of-the-art self-attention models for machine translation and image super-resolution. Understand how the proposed technique achieves iteration reductions of up to 2x over baseline greedy decoders without quality loss, or up to 7x with a slight performance decrease. Examine the real-time speedups of up to 4x over standard greedy decoding in terms of wall-clock time. Gain insights into the trade-offs between computation needed per layer and critical path length at training time for different architecture classes such as recurrent, convolutional, and self-attention networks.

Syllabus

Blockwise Parallel Decoding for Deep Autoregressive Models


Taught by

Yannic Kilcher

Related Courses

Introduction to Artificial Intelligence
Stanford University via Udacity
Computer Vision: The Fundamentals
University of California, Berkeley via Coursera
Computational Photography
Georgia Institute of Technology via Coursera
Digital Signal Processing
École Polytechnique Fédérale de Lausanne via Coursera
Creative, Serious and Playful Science of Android Apps
University of Illinois at Urbana-Champaign via Coursera