Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Recurrent Neural Networks, Transformers, and Attention

Alexander Amini via YouTube

Overview

Dive into the world of advanced deep learning techniques with this comprehensive lecture from MIT's Introduction to Deep Learning course. Explore the intricacies of Recurrent Neural Networks (RNNs), Transformers, and Attention mechanisms. Begin with an introduction to sequence modeling and neurons with recurrence, then delve into the fundamentals of RNNs, including their intuition and unfolding process. Learn how to build RNNs from scratch and understand the design criteria for sequential modeling through a word prediction example. Discover the backpropagation through time algorithm and address gradient issues in RNNs. Investigate Long Short-Term Memory (LSTM) networks and their applications. Finally, uncover the power of attention mechanisms, their intuition, relationship to search, and implementation in neural networks. Gain insights into scaling attention and its various applications in deep learning.

Syllabus

​ - Introduction
​ - Sequence modeling
​ - Neurons with recurrence
- Recurrent neural networks
- RNN intuition
​ - Unfolding RNNs
- RNNs from scratch
- Design criteria for sequential modeling
- Word prediction example
​ - Backpropagation through time
- Gradient issues
​ - Long short term memory LSTM
​ - RNN applications
- Attention fundamentals
- Intuition of attention
- Attention and search relationship
- Learning attention with neural networks
- Scaling attention and applications
- Summary

Taught by

https://www.youtube.com/@AAmini/videos

Reviews

Start your review of Recurrent Neural Networks, Transformers, and Attention

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.