Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Enabling Efficient Trillion Parameter Scale Training for Deep Learning Models

MLOps.community via YouTube

Overview

Explore the challenges and solutions for efficient trillion-parameter scale training in deep learning models in this conference talk from the AI in Production Conference. Delve into DeepSpeed, a deep learning optimization library designed to make distributed model training and inference efficient, effective, and easy on commodity hardware. Learn about training optimizations that improve memory, compute, and data efficiency for extreme model scaling. Gain insights from Olatunji (Tunji) Ruwase, co-founder and lead of the DeepSpeed project at Microsoft, as he shares his expertise in building systems convergence optimizations and frameworks for distributed training and inference of deep learning models.

Syllabus

Enabling Efficient Trillion Parameter Scale Training for Deep Learning Models // Tunji Ruwase

Taught by

MLOps.community

Reviews

Start your review of Enabling Efficient Trillion Parameter Scale Training for Deep Learning Models

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.