Overview
Watch a 46-minute research talk from Google DeepMind's Andreas Terzis exploring the intersection of differential privacy and synthetic data generation for training large language models. Delve into critical aspects of LLM development including alignment, trust mechanisms, watermarking techniques, and copyright considerations while learning about innovative approaches to maintain privacy in AI training data. Gain insights into how synthetic data can be generated with differential privacy guarantees to protect sensitive information while still producing effective training datasets for language models.
Syllabus
Differentially private synthetic data for private LLM training
Taught by
Simons Institute