Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

RecurrentGemma: Moving Past Transformers with Griffin Architecture for Long Context Length

Discover AI via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a comprehensive technical video that delves into Google's groundbreaking RecurrentLLM architecture with Griffin, presenting a significant shift from traditional transformer-based models. Learn about the innovative RecurrentGemma-2B model, which achieves an impressive throughput of 6000 tokens per second while maintaining performance comparable to transformer-based Gemma 2B. Discover the technical intricacies of new architectures like GRIFFIN and HAWK, with detailed explanations of their advantages over State Space Models such as Mamba-S6. Master concepts including local attention mechanisms, linear recurrences, GRU (Gated Recurrent Unit), LRU (Linear Recurrent Unit), and RG-LRU (Real-Gated Linear Recurrent Unit). Gain insights into the model's fixed-size state architecture, which offers superior memory efficiency for long sequences compared to traditional transformer models' growing key-value cache. Examine performance benchmarks, practical implementations through Github code examples, and understand how this architectural innovation maintains high throughput regardless of sequence length while requiring 33% fewer training tokens than its transformer counterpart.

Syllabus

Llama 3 inference and finetuning
New Language Model Dev
Local Attention
Linear complexity of RNN
Gated recurrent unit - GRU
Linear recurrent Unit - LRU
GRIFFIN architecture
Real-Gated Linear recurrent unit RG-LRU
Griffin Key Features
RecurrentGemma
Github code
Performance benchmark

Taught by

Discover AI

Reviews

Start your review of RecurrentGemma: Moving Past Transformers with Griffin Architecture for Long Context Length

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.