Overview
Explore the concept of essential sparsity in large pre-trained models through this insightful 1-hour 10-minute talk by Professor Atwas Wang from the University of Austin Texas. Delve into efficient methods for handling complex and expansive pre-trained transformer models in contemporary machine learning. Discover the threshold at which removing small-magnitude weights significantly impacts performance compared to lower levels of sparsity. Gain access to the project code on GitHub and learn about the research paper "The Emergence of Essential Sparsity in Large Pre-trained Models: The Weights that Matter." Explore additional resources such as The Deep Dive newsletter for the latest AI research and industry trends, and Unify's blog for insights into the AI deployment stack. Connect with Unify through their website, GitHub, Discord, and Twitter to stay updated on AI advancements, transformers, language models, and sparsification techniques.
Syllabus
The Emergence of Essential Sparsity in Large Pre-trained Models
Taught by
Unify