GLIDE- Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models
Offered By: Yannic Kilcher via YouTube
Course Description
Overview
Explore the groundbreaking GLIDE model for text-to-image generation in this comprehensive video lecture. Delve into the mechanics of diffusion models and their application in creating photorealistic images from text descriptions. Learn about conditional generation techniques, guided diffusion, and the architecture behind GLIDE. Examine training methodologies, result metrics, and potential failure cases. Gain insights into safety considerations surrounding this powerful technology. Discover how GLIDE compares to other models like DALL-E and understand its implications for text-driven image editing and inpainting.
Syllabus
- Intro & Overview
- What is a Diffusion Model?
- Conditional Generation and Guided Diffusion
- Architecture Recap
- Training & Result metrics
- Failure cases & my own results
- Safety considerations
Taught by
Yannic Kilcher
Related Courses
Diffusion Models Beat GANs on Image Synthesis - Machine Learning Research Paper ExplainedYannic Kilcher via YouTube Diffusion Models Beat GANs on Image Synthesis - ML Coding Series - Part 2
Aleksa Gordić - The AI Epiphany via YouTube OpenAI GLIDE - Towards Photorealistic Image Generation and Editing with Text-Guided Diffusion Models
Aleksa Gordić - The AI Epiphany via YouTube Food for Diffusion
HuggingFace via YouTube Imagen: Text-to-Image Generation Using Diffusion Models - Lecture 9
University of Central Florida via YouTube