Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Scalable Extraction of Training Data from Language Models

Yannic Kilcher via YouTube

Overview

Explore a detailed analysis of a research paper revealing how large language models like ChatGPT can inadvertently leak training data through simple prompts. Delve into the concepts of extractable and discoverable memorization, examining how researchers were able to extract gigabytes of data from various models. Learn about the novel "divergence attack" developed to exploit ChatGPT, causing it to deviate from normal behavior and emit training data at a much higher rate. Understand the implications of these findings for data privacy, model security, and the effectiveness of current alignment techniques in preventing memorization. Gain insights into quantitative membership testing and the broader consequences of this research for the field of AI and machine learning.

Syllabus

- Intro
- Extractable vs Discoverable Memorization
- Models leak more data than previously thought
- Some data is extractable but not discoverable
- Extracting data from closed models
- Poem poem poem
- Quantitative membership testing
- Exploring the ChatGPT exploit further
- Conclusion

Taught by

Yannic Kilcher

Reviews

Start your review of Scalable Extraction of Training Data from Language Models

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.