Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

MobileLLM - Optimizing Sub-billion Parameter Language Models for On-Device Use Cases

EDGE AI FOUNDATION via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Watch a 20-minute research presentation exploring the development of MobileLLM, a groundbreaking approach to deploying efficient large language models on mobile devices. Learn how deep and thin architectures, embedding sharing, and grouped-query attention mechanisms enable high-performance language models with fewer than a billion parameters. Discover how these optimizations achieve significant accuracy improvements over previous state-of-the-art models in commonsense reasoning tasks, with 2.7% and 4.3% boosts for 125M and 350M parameter models respectively. Understand how this architectural innovation challenges the conventional wisdom that data and parameter quantity are the primary drivers of model quality, while demonstrating comparable performance to much larger models in practical applications like API calling tasks.

Syllabus

GenAI on the Edge Forum: MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device..

Taught by

EDGE AI FOUNDATION

Reviews

Start your review of MobileLLM - Optimizing Sub-billion Parameter Language Models for On-Device Use Cases

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.