YoVDO

OpenAI CLIP- Connecting Text and Images

Offered By: Yannic Kilcher via YouTube

Tags

Computer Vision Courses Artificial Intelligence Courses

Course Description

Overview

Explore a comprehensive video explanation of OpenAI's CLIP (Contrastive Language-Image Pre-training) model, which connects text and images. Delve into the paper "Learning Transferable Visual Models From Natural Language Supervision" and understand how CLIP trains on 400 million web-scraped images with text descriptions to create a versatile model. Learn about the contrastive objective, large batch size implementation, and how the resulting model can be adapted for zero-shot classification tasks. Examine the model's architecture, training process, performance comparisons, scaling properties, and robustness to data shift. Gain insights into the broader impact of this technology and its potential applications in various computer vision tasks.

Syllabus

- Introduction
- Overview
- Connecting Images & Text
- Building Zero-Shot Classifiers
- CLIP Contrastive Training Objective
- Encoder Choices
- Zero-Shot CLIP vs Linear ResNet-50
- Zero-Shot vs Few-Shot
- Scaling Properties
- Comparison on different tasks
- Robustness to Data Shift
- Broader Impact Section
- Conclusion & Comments


Taught by

Yannic Kilcher

Related Courses

Introduction to Artificial Intelligence
Stanford University via Udacity
Probabilistic Graphical Models 1: Representation
Stanford University via Coursera
Artificial Intelligence for Robotics
Stanford University via Udacity
Computer Vision: The Fundamentals
University of California, Berkeley via Coursera
Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent