TransGAN - Two Transformers Can Make One Strong GAN - Machine Learning Research Paper Explained
Offered By: Yannic Kilcher via YouTube
Course Description
Overview
Explore a comprehensive video explanation of the machine learning research paper "TransGAN: Two Transformers Can Make One Strong GAN." Delve into the groundbreaking approach of using transformer-based architectures for both the generator and discriminator in Generative Adversarial Networks (GANs). Learn about the innovative techniques employed, including data augmentation with DiffAug, super-resolution co-training, and localized initialization of self-attention. Discover how TransGAN achieves competitive performance with convolutional GANs on various datasets and gain insights into the future potential of transformer-based GANs in computer vision tasks.
Syllabus
- Introduction & Overview
- Discriminator Architecture
- Generator Architecture
- Upsampling with PixelShuffle
- Architecture Recap
- Vanilla TransGAN Results
- Trick 1: Data Augmentation with DiffAugment
- Trick 2: Super-Resolution Co-Training
- Trick 3: Locality-Aware Initialization for Self-Attention
- Scaling Up & Experimental Results
- Recap & Conclusion
Taught by
Yannic Kilcher
Related Courses
TensorFlow を使った畳み込みニューラルネットワークDeepLearning.AI via Coursera Emotion AI: Facial Key-points Detection
Coursera Project Network via Coursera Transfer Learning for Food Classification
Coursera Project Network via Coursera Facial Expression Classification Using Residual Neural Nets
Coursera Project Network via Coursera Apply Generative Adversarial Networks (GANs)
DeepLearning.AI via Coursera