Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the phenomenon of 'benign overfitting' in deep learning through this seminar by Peter Bartlett, Associate Director of the Simons Institute for Theory of Computing and professor at UC Berkeley. Delve into the intriguing world of machine learning where traditional theories of balancing data fit and rule complexity are challenged by deep networks' ability to perform well despite perfectly fitting noisy training data. Examine this concept in the context of linear prediction, uncovering a characterization of regression problems where minimum norm interpolating prediction rules achieve near-optimal accuracy. Discover the importance of massive overparameterization and its implications for deep networks and robustness to adversarial examples. Gain insights from Bartlett's extensive research in statistical learning theory and his contributions to the field of machine learning.