Overview
Syllabus
Intro
The ERM/SRM theory of learning
Unifom laws of large numbers
Capacity control
U-shaped generalization curve
Does interpolation overfit?
Interpolation does not averfit even for very noisy data
why bounds fail
Interpolation is best practice for deep learning
Historical recognition
The key lesson
Generalization theory for interpolation?
A way forward?
Interpolated k-NN schemes
Interpolation and adversarial examples
Double descent risk curve
More parameters are better: an example
Random Fourier networks
what is the mechanism?
Double Descent in Randon Feature settings
Smoothness by averaging
Framework for modern ML
The landscape of generalization
Optimization: classical
Modern Optimization
From classical statistics to modern ML
The nature of inductive bias
Memorization and interpolation
Interpolation in deep auto-encoders
Neural networks as models for associative memory
Why are attractors surprising?
Memorizing sequences
Taught by
MITCBMM