Completed
Constructing Self Attention Matrix
Class Central Classrooms beta
YouTube videos curated by Class Central.
Classroom Contents
Deep Dive into the Transformer Encoder Architecture
Automatically move to the next video in the Classroom when playback concludes
- 1 Introduction
- 2 Encoder Overview
- 3 Blowing up the encoder
- 4 Create Initial Embeddings
- 5 Positional Encodings
- 6 The Encoder Layer Begins
- 7 Query, Key, Value Vectors
- 8 Constructing Self Attention Matrix
- 9 Why scaling and Softmax?
- 10 Combining Attention heads
- 11 Residual Connections Skip Connections
- 12 Layer Normalization
- 13 Why Linear Layers, ReLU, Dropout
- 14 Complete the Encoder Layer
- 15 Final Word Embeddings
- 16 Sneak Peak of Code