Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Self - Cross, Hard - Soft Attention and the Transformer

Alfredo Canziani via YouTube

Overview

Explore the intricacies of attention mechanisms and Transformer architecture in this comprehensive lecture. Delve into self-attention, cross-attention, hard attention, and soft attention concepts. Learn about set encoding use cases and the key-value store paradigm. Understand the implementation of queries, keys, and values in both self-attention and cross-attention contexts. Examine the Transformer's encoder-predictor-decoder architecture, with a focus on the encoder and the unique "decoder" module. Gain practical insights through a PyTorch implementation of a Transformer encoder using Jupyter Notebook. Additionally, discover useful tips for reading and summarizing research papers collaboratively.

Syllabus

– Welcome to class
– Listening to YouTube from the terminal
– Summarising papers with @Notion
– Reading papers collaboratively
– Attention! Self / cross, hard / soft
– Use cases: set encoding!
– Self-attention
– Key-value store
– Queries, keys, and values → self-attention
– Queries, keys, and values → cross-attention
– Implementation details
– The Transformer: an encoder-predictor-decoder architecture
– The Transformer encoder
– The Transformer “decoder” which is an encoder-predictor-decoder module
– Jupyter Notebook and PyTorch implementation of a Transformer encoder
– Goodbye :

Taught by

Alfredo Canziani

Reviews

Start your review of Self - Cross, Hard - Soft Attention and the Transformer

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.