Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

OpenAI CLIP- Connecting Text and Images

Yannic Kilcher via YouTube

Overview

Explore a comprehensive video explanation of OpenAI's CLIP (Contrastive Language-Image Pre-training) model, which connects text and images. Delve into the paper "Learning Transferable Visual Models From Natural Language Supervision" and understand how CLIP trains on 400 million web-scraped images with text descriptions to create a versatile model. Learn about the contrastive objective, large batch size implementation, and how the resulting model can be adapted for zero-shot classification tasks. Examine the model's architecture, training process, performance comparisons, scaling properties, and robustness to data shift. Gain insights into the broader impact of this technology and its potential applications in various computer vision tasks.

Syllabus

- Introduction
- Overview
- Connecting Images & Text
- Building Zero-Shot Classifiers
- CLIP Contrastive Training Objective
- Encoder Choices
- Zero-Shot CLIP vs Linear ResNet-50
- Zero-Shot vs Few-Shot
- Scaling Properties
- Comparison on different tasks
- Robustness to Data Shift
- Broader Impact Section
- Conclusion & Comments

Taught by

Yannic Kilcher

Reviews

Start your review of OpenAI CLIP- Connecting Text and Images

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.