Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Recent Developments in Over-parametrized Neural Networks, Part I

Simons Institute via YouTube

Overview

Explore recent developments in over-parametrized neural networks in this comprehensive lecture from the Deep Learning Boot Camp. Delve into topics such as non-convex optimization, overparametrization, and theoretical problems in optimal optimization. Examine the taxonomy of results, including skip connections, expressivity, and geometric results. Investigate one-point convexity, Taylor's theorem, and the Royer algorithm. Learn about second-order stationary points, optimality conditions, and global optimality in neural networks. Discover strategies for second-order descent and stationary points in learning networks. Gain valuable insights from Jason Lee of the University of Southern California as he presents cutting-edge research and concepts in the field of deep learning and neural networks.

Syllabus

Introduction
Outline
Goals
Whats Hard
Whats Easy
Non convex optimization
Notation
Overparametrization
Experiments
Rule of Thumb
Theoretical Problem
Optimal Optimization
Taxonomy of Results
Skip Connections
Expressivity
Geometric Results
Geometry Results
Onepoint convexity
Taylors theorem
Royer algorithm
Randomness
Theorem
Secondorder Stationary Points
Neural Networks
Optimality Conditions
Global Optimal
Other Losses
NonDegenerate Critical Points
Three Strategies
Second Order Descent
Stationary Points
Learning Networks

Taught by

Simons Institute

Reviews

Start your review of Recent Developments in Over-parametrized Neural Networks, Part I

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.