Overview
Syllabus
Intro
What might generalization theory offer deep learning?
Barriers to explaining generalization
PAC-Bayes yields risk bounds for Gibbs classifiers
PAC-Bayes generalization bounds
PAC-Bayes bounds on deterministic classifiers
Distribution-dependent approximations of optimal priors via privacy
A question of interpretation
Use SGD to predict SGD
Data and distribution priors for neural networks
MNIST Results - Coupled data dependent priors and posteriors
Oracle access to optimal prior covariance
Bounds with oracle covariance + ghost sample
Bounds on 32k samples v 64k samples
Recap and Conclusion
Taught by
Simons Institute