Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Matrix Long Short-Term Memory (mLSTM) - A New Alternative to Transformer LLMs

Discover AI via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a detailed technical analysis of the newly published xLSTM architecture, specifically focusing on the Matrix Long Short-Term Memory (mLSTM) network, in this 23-minute video presentation. Dive into the innovative concept of "accumulated covariance" with exponential gating functions and understand how this advanced variation of traditional LSTM models compares to classical attention mechanisms. Learn about the matrix-based approach that differentiates mLSTM, where input and recurrent weights along with gates are represented as matrices instead of vectors, enabling more sophisticated data processing. Discover how this architecture enhances the network's ability to capture complex relationships and dependencies within data through matrix operations, potentially offering improved representational power and computational efficiency for natural language processing and time series analysis tasks. While independent performance evaluation is pending due to the recent publication, gain valuable insights into this potential alternative to transformer LLMs and its theoretical advantages in handling high-dimensional datasets.

Syllabus

New xLSTM explained: Better than Transformer LLMs?

Taught by

Discover AI

Reviews

Start your review of Matrix Long Short-Term Memory (mLSTM) - A New Alternative to Transformer LLMs

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.