Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Model Stealing for Low-Rank Language Models Through Reverse Engineering

Discover AI via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Learn about groundbreaking research in AI cybersecurity and data privacy through a 28-minute technical video exploring methods for reverse-engineering Large Language Models (LLMs). Delve into MIT's innovative approach for efficiently learning and sampling from low-rank distributions over sequences, featuring detailed explanations of Hidden Markov Models, barycentric spanners, and convex optimization techniques. Master the mathematical foundations behind a novel method that uses conditional queries and dimensionality reduction to reconstruct transition models and generate sequences mimicking LLM behavior. Follow along as MIT researchers demonstrate how to capture essential features of complex language models without requiring access to their parameters or training data. Progress through key concepts including KL divergence, low-rank distributions, and the mathematical theorems underpinning this breakthrough in AI model analysis.

Syllabus

Model Stealing for ANY Low Rank Language Model
Learning Hidden Markov Models
Reverse-Engineer LLMs
Professor of Mathematics MIT
Hidden Markov Models explained
New method
Barycentric Spanner explained
Convex Optimization KL Divergence
Low Rank Distribution explained
MAIN Challenge
The MAIN Mathematical Theorem

Taught by

Discover AI

Reviews

Start your review of Model Stealing for Low-Rank Language Models Through Reverse Engineering

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.