Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

RoPE: Rotary Position Embedding for Extended Context Lengths in Transformers

Discover AI via YouTube

Overview

Learn about Rotary Position Embedding (RoPE) in this 40-minute technical video that breaks down complex concepts into simple terms for understanding how self-attention works in Transformers with relative position encoding. Explore the mathematical foundations and practical applications of RoPE that enable Large Language Models (LLMs) to handle extended context lengths up to 100K tokens. Dive into the key concepts from the RoFormer paper, examining how rotary position embeddings enhance transformer architectures for improved performance in natural language processing tasks. Gain valuable insights into this advanced AI research topic through clear explanations and detailed breakdowns of the underlying mechanisms.

Syllabus

RoPE Rotary Position Embedding to 100K context length

Taught by

Discover AI

Reviews

Start your review of RoPE: Rotary Position Embedding for Extended Context Lengths in Transformers

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.