Mamba: Linear-Time Sequence Modeling with Selective State Spaces

Mamba: Linear-Time Sequence Modeling with Selective State Spaces

Yannic Kilcher via YouTube Direct link

- Efficient computation via prefix sums / parallel scans

8 of 10

8 of 10

- Efficient computation via prefix sums / parallel scans

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Mamba: Linear-Time Sequence Modeling with Selective State Spaces

Automatically move to the next video in the Classroom when playback concludes

  1. 1 - Introduction
  2. 2 - Transformers vs RNNs vs S4
  3. 3 - What are state space models?
  4. 4 - Selective State Space Models
  5. 5 - The Mamba architecture
  6. 6 - The SSM layer and forward propagation
  7. 7 - Utilizing GPU memory hierarchy
  8. 8 - Efficient computation via prefix sums / parallel scans
  9. 9 - Experimental results and comments
  10. 10 - A brief look at the code

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.