YoVDO

DALL-E - Zero-Shot Text-to-Image Generation - Paper Explained

Offered By: Aleksa Gordić - The AI Epiphany via YouTube

Tags

DALL-E Courses Machine Learning Courses Zero-shot learning (ZSL) Courses Generative Modeling Courses

Course Description

Overview

Dive into a comprehensive video explanation of OpenAI's DALL-E paper on zero-shot text-to-image generation. Explore the two-stage process involving VQ-VAE and autoregressive transformers, understand ELBO concepts, and discover how the model combines distinct concepts to create plausible images. Learn about engineering challenges, automatic filtering using CLIP, and witness impressive results including image-to-image translation capabilities. Gain insights into this groundbreaking AI technology through detailed explanations and visual examples.

Syllabus

What is DALL-E?
VQ-VAE blur problems
transformers, transformers, transformers!
Stage 1 and Stage 2 explained
Stage 1 VQ-VAE recap
Stage 2 autoregressive transformer
Some notes on ELBO
VQ-VAE modifications
Stage 2 in-depth
Results
Engineering, engineering, engineering
Automatic filtering via CLIP
More results
Additional image to image translation examples


Taught by

Aleksa Gordić - The AI Epiphany

Related Courses

Fast Zero Shot Object Detection with OpenAI CLIP
James Briggs via YouTube
Language Models as Zero-Shot Planners - Extracting Actionable Knowledge for Embodied Agents
Yannic Kilcher via YouTube
Tensor Programs - Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer
Aleksa Gordić - The AI Epiphany via YouTube
Tuning GPT-3 on a Single GPU via Zero-Shot Hyperparameter Transfer
Massachusetts Institute of Technology via YouTube