Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

How to Steal ChatGPT's Embedding Size and Other Low-rank Logit Tricks

USC Information Sciences Institute via YouTube

Overview

Explore the implications of large language model (LLM) commercialization and API restrictions in this 48-minute talk presented by Matt Finlayson from USC Information Sciences Institute. Discover how, with minimal assumptions about model architecture, significant non-public information can be extracted from API-protected LLMs using a relatively small number of queries. Learn about the softmax bottleneck in modern LLMs and how it can be exploited to obtain full-vocabulary outputs, audit model updates, identify source LLMs, and even uncover hidden model sizes. Examine the empirical investigations that led to estimating OpenAI's gpt-3.5-turbo embedding size at approximately 4096. Consider potential safeguards against these techniques and discuss how these capabilities might contribute to greater transparency and accountability in AI development. Gain insights from Finlayson's background in NLP, computer science, and linguistics as he explores the practical consequences of language model architectural design, from security to generation and learning processes.

Syllabus

How to Steal ChatGPT’s Embedding Size, and Other Low-rank Logit Tricks

Taught by

USC Information Sciences Institute

Reviews

Start your review of How to Steal ChatGPT's Embedding Size and Other Low-rank Logit Tricks

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.