AttnGAN: Fine-Grained Text to Image Generation with Attentional Generative Adversarial Networks
Offered By: University of Central Florida via YouTube
Course Description
Overview
Explore the innovative AttnGAN model for fine-grained text-to-image generation in this 46-minute lecture from the University of Central Florida. Delve into the architecture's key components, including the text encoder, conditioning augmentation, generator, attention network, and image encoder. Examine the DAMSM loss and its role in improving image quality. Learn about experimental results on various datasets, evaluation metrics like Inception score, and component analysis. Discover the model's capabilities in generating novel scenarios and understand its limitations in capturing global coherent structure. Gain insights into the challenges and advancements in text-to-image synthesis using attentional generative adversarial networks.
Syllabus
Intro
Problem: Text-to-image
Related work
Architecture - Motivation
Architecture - Text Encoder
Architecture - Conditioning Augmentation
Architecture - Generator F.
Architecture - Attention network Fatin
Architecture - Image Encoder
Architecture - DAMSM loss
Experiments - Datasets
Experiments - Evaluation • Inception score
Experiments - Component Analysis
Experiments - Qualitative (CUB)
Experiments - Novel scenarios
Experiments - Failure cases Did not capture global coherent structure
Taught by
UCF CRCV
Tags
Related Courses
2D image processingHigher School of Economics via Coursera 3D Reconstruction - Multiple Viewpoints
Columbia University via Coursera 3D Reconstruction - Single Viewpoint
Columbia University via Coursera AI-900: Microsoft Certified Azure AI Fundamentals
A Cloud Guru TensorFlow Developer Certificate Exam Prep
A Cloud Guru