Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Distributed Training Methods for Efficient Machine Learning - Part 1

MIT HAN Lab via YouTube

Overview

Learn about distributed training fundamentals in machine learning through a comprehensive MIT lecture that explores parallelization methods, data parallelism, and communication primitives. Dive deep into memory reduction techniques like ZeRO and FSDP, while understanding pipeline parallelism, tensor parallelism, and sequence parallelism. Professor Song Han delivers this 70-minute lecture covering essential background, motivation, and various parallelization approaches for training large-scale machine learning models efficiently across distributed systems.

Syllabus

EfficientML.ai Lecture 19 - Distributed Training Part 1 (MIT 6.5940, Fall 2024)

Taught by

MIT HAN Lab

Reviews

Start your review of Distributed Training Methods for Efficient Machine Learning - Part 1

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.