RWKV- Reinventing RNNs for the Transformer Era

RWKV- Reinventing RNNs for the Transformer Era

Yannic Kilcher via YouTube Direct link

- RWKV's Layer Structure

7 of 11

7 of 11

- RWKV's Layer Structure

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

RWKV- Reinventing RNNs for the Transformer Era

Automatically move to the next video in the Classroom when playback concludes

  1. 1 - Introduction
  2. 2 - Fully Connected In-Person Conference in SF June 7th
  3. 3 - Transformers vs RNNs
  4. 4 - RWKV: Best of both worlds
  5. 5 - LSTMs
  6. 6 - Evolution of RWKV's Linear Attention
  7. 7 - RWKV's Layer Structure
  8. 8 - Time-Parallel vs Sequence Mode
  9. 9 - Experimental Results & Limitations
  10. 10 - Visualizations
  11. 11 - Conclusion

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.