Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Gradient Descent and the Backpropagation Algorithm

Alfredo Canziani via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Dive into a comprehensive lecture on gradient descent and the backpropagation algorithm, delivered by renowned speaker Yann LeCun. Explore key concepts in supervised learning, parametrized models, and loss functions before delving into the intricacies of gradient descent. Gain insights into traditional neural networks and learn how backpropagation works through non-linear functions and weighted sums. Follow along with a PyTorch implementation and discover practical applications of backpropagation. Investigate the process of learning representations and understand why shallow networks are considered universal approximators. Conclude by examining the relationship between multilayer architectures and the compositional structure of data in this in-depth, 1-hour and 51-minute exploration of fundamental machine learning concepts.

Syllabus

– Supervised learning
– Parametrised models
– Block diagram
– Loss function, average loss
– Gradient descent
– Traditional neural nets
– Backprop through a non-linear function
– Backprop through a weighted sum
– PyTorch implementation
– Backprop through a functional module
– Backprop through a functional module
– Backprop in practice
– Learning representations
– Shallow networks are universal approximators!
– Multilayer architectures == compositional structure of data

Taught by

Alfredo Canziani

Reviews

Start your review of Gradient Descent and the Backpropagation Algorithm

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.