Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Large Model Training and Inference with DeepSpeed

MLOps.community via YouTube

Overview

Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the journey of DeepSpeed and its transformative impact on large model training and inference in this 36-minute conference talk by Samyam Rajbhandari at the LLMs in Prod Conference. Discover how technologies like ZeRO and 3D-Parallelism have become fundamental building blocks for training large language models at scale, powering LLMs such as Bloom-176B and Megatron-Turing 530B. Learn about heterogeneous memory training systems like ZeRO-Offload and ZeRO-Infinity, which have democratized LLMs by making them accessible with limited resources. Gain insights into DeepSpeed-Inference and DeepSpeed-MII, which simplify the application of powerful inference optimizations to accelerate LLMs for deployment. Understand how DeepSpeed has been integrated into platforms like HuggingFace, PyTorch Lightning, and Mosaic ML, and how its technologies are offered in PyTorch, Colossal-AI, and Megatron-LM. Delve into the motivations, insights, and stories behind the development of these groundbreaking technologies that have revolutionized large language model training and inference.

Syllabus

Large Model Training and Inference with DeepSpeed // Samyam Rajbhandari // LLMs in Prod Conference

Taught by

MLOps.community

Reviews

Start your review of Large Model Training and Inference with DeepSpeed

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.