OpenAI CLIP- Connecting Text and Images
Offered By: Yannic Kilcher via YouTube
Course Description
Overview
Explore a comprehensive video explanation of OpenAI's CLIP (Contrastive Language-Image Pre-training) model, which connects text and images. Delve into the paper "Learning Transferable Visual Models From Natural Language Supervision" and understand how CLIP trains on 400 million web-scraped images with text descriptions to create a versatile model. Learn about the contrastive objective, large batch size implementation, and how the resulting model can be adapted for zero-shot classification tasks. Examine the model's architecture, training process, performance comparisons, scaling properties, and robustness to data shift. Gain insights into the broader impact of this technology and its potential applications in various computer vision tasks.
Syllabus
- Introduction
- Overview
- Connecting Images & Text
- Building Zero-Shot Classifiers
- CLIP Contrastive Training Objective
- Encoder Choices
- Zero-Shot CLIP vs Linear ResNet-50
- Zero-Shot vs Few-Shot
- Scaling Properties
- Comparison on different tasks
- Robustness to Data Shift
- Broader Impact Section
- Conclusion & Comments
Taught by
Yannic Kilcher
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Artificial Intelligence for Robotics
Stanford University via Udacity Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent