Completed
- The Complexity of Self-Attention
Class Central Classrooms beta
YouTube videos curated by Class Central.
Classroom Contents
Linformer - Self-Attention with Linear Complexity
Automatically move to the next video in the Classroom when playback concludes
- 1 - Intro & Overview
- 2 - The Complexity of Self-Attention
- 3 - Embedding Dimension & Multiple Heads
- 4 - Formal Attention
- 5 - Empirical Investigation into RoBERTa
- 6 - Theorem: Self-Attention is Low Rank
- 7 - Linear Self-Attention Method
- 8 - Theorem: Linear Self-Attention
- 9 - Language Modeling
- 10 - NLP Benchmarks
- 11 - Compute Time & Memory Gains
- 12 - Broader Impact Statement
- 13 - Conclusion