Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a comprehensive video explanation of OpenAI's CLIP (Contrastive Language-Image Pre-training) model, which connects text and images. Delve into the paper "Learning Transferable Visual Models From Natural Language Supervision" and understand how CLIP trains on 400 million web-scraped images with text descriptions to create a versatile model. Learn about the contrastive objective, large batch size implementation, and how the resulting model can be adapted for zero-shot classification tasks. Examine the model's architecture, training process, performance comparisons, scaling properties, and robustness to data shift. Gain insights into the broader impact of this technology and its potential applications in various computer vision tasks.
Syllabus
- Introduction
- Overview
- Connecting Images & Text
- Building Zero-Shot Classifiers
- CLIP Contrastive Training Objective
- Encoder Choices
- Zero-Shot CLIP vs Linear ResNet-50
- Zero-Shot vs Few-Shot
- Scaling Properties
- Comparison on different tasks
- Robustness to Data Shift
- Broader Impact Section
- Conclusion & Comments
Taught by
Yannic Kilcher