Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

The Emergence of Essential Sparsity in Large Pre-trained Models

Unify via YouTube

Overview

Explore the concept of essential sparsity in large pre-trained models through this insightful 1-hour 10-minute talk by Professor Atwas Wang from the University of Austin Texas. Delve into efficient methods for handling complex and expansive pre-trained transformer models in contemporary machine learning. Discover the threshold at which removing small-magnitude weights significantly impacts performance compared to lower levels of sparsity. Gain access to the project code on GitHub and learn about the research paper "The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter." Explore additional resources such as The Deep Dive newsletter for the latest AI research and industry trends, and Unify's blog for insights into the AI deployment stack. Connect with Unify through their website, GitHub, Discord, and Twitter to stay updated on AI advancements, transformers, language models, and sparsification techniques.

Syllabus

The Emergence of Essential Sparsity in Large Pre-trained Models

Taught by

Unify

Reviews

Start your review of The Emergence of Essential Sparsity in Large Pre-trained Models

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.