Overview
Syllabus
Intro
Outline
Starting point
Single-hidden layer (shallow) neural networks of infinite width Consider a NN which
What GP does it correspond to?
Properties of the NNGP
Bayesian inference with a GP prior (Review)
Experiments from original work
Performance comparison
NNGP performance across hyperparameters
Large depth behavior & fixed points
Phase diagrams: experiments vs. theory
Performance trends with width and dataset size
Empirical comparison of various NN-GPS
Empirical trends
Best performing networks: comparison between GPs and SGD-NNS
Partway summary
What dynamics occurs in parameter space?
Closing Remarks
Taught by
Simons Institute