Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Learn about groundbreaking research in AI cybersecurity and data privacy through a 28-minute technical video exploring methods for reverse-engineering Large Language Models (LLMs). Delve into MIT's innovative approach for efficiently learning and sampling from low-rank distributions over sequences, featuring detailed explanations of Hidden Markov Models, barycentric spanners, and convex optimization techniques. Master the mathematical foundations behind a novel method that uses conditional queries and dimensionality reduction to reconstruct transition models and generate sequences mimicking LLM behavior. Follow along as MIT researchers demonstrate how to capture essential features of complex language models without requiring access to their parameters or training data. Progress through key concepts including KL divergence, low-rank distributions, and the mathematical theorems underpinning this breakthrough in AI model analysis.
Syllabus
Model Stealing for ANY Low Rank Language Model
Learning Hidden Markov Models
Reverse-Engineer LLMs
Professor of Mathematics MIT
Hidden Markov Models explained
New method
Barycentric Spanner explained
Convex Optimization KL Divergence
Low Rank Distribution explained
MAIN Challenge
The MAIN Mathematical Theorem
Taught by
Discover AI