DALL-E - Zero-Shot Text-to-Image Generation - Paper Explained
Offered By: Aleksa Gordić - The AI Epiphany via YouTube
Course Description
Overview
Dive into a comprehensive video explanation of OpenAI's DALL-E paper on zero-shot text-to-image generation. Explore the two-stage process involving VQ-VAE and autoregressive transformers, understand ELBO concepts, and discover how the model combines distinct concepts to create plausible images. Learn about engineering challenges, automatic filtering using CLIP, and witness impressive results including image-to-image translation capabilities. Gain insights into this groundbreaking AI technology through detailed explanations and visual examples.
Syllabus
What is DALL-E?
VQ-VAE blur problems
transformers, transformers, transformers!
Stage 1 and Stage 2 explained
Stage 1 VQ-VAE recap
Stage 2 autoregressive transformer
Some notes on ELBO
VQ-VAE modifications
Stage 2 in-depth
Results
Engineering, engineering, engineering
Automatic filtering via CLIP
More results
Additional image to image translation examples
Taught by
Aleksa Gordić - The AI Epiphany
Related Courses
Fast Zero Shot Object Detection with OpenAI CLIPJames Briggs via YouTube Language Models as Zero-Shot Planners - Extracting Actionable Knowledge for Embodied Agents
Yannic Kilcher via YouTube Tensor Programs - Tuning Large Neural Networks via Zero-Shot Hyperparameter Transfer
Aleksa Gordić - The AI Epiphany via YouTube Tuning GPT-3 on a Single GPU via Zero-Shot Hyperparameter Transfer
Massachusetts Institute of Technology via YouTube