Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Distributed Training and Gradient Compression - Part I - Lecture 13

MIT HAN Lab via YouTube

Overview

Explore the fundamentals of distributed training for neural networks in this lecture from MIT's course on TinyML and Efficient Deep Learning Computing. Delve into key concepts such as data parallelism and model parallelism, essential for scaling machine learning models across multiple devices. Learn how to overcome challenges in deploying neural networks on mobile and IoT devices, and discover techniques for accelerating training processes. Gain insights from instructor Song Han on efficient machine learning methods that enable powerful deep learning applications on resource-constrained devices. Access accompanying slides and course materials to enhance your understanding of distributed training strategies and their practical applications in mobile AI and IoT scenarios.

Syllabus

Lecture 13 - Distributed Training and Gradient Compression (Part I) | MIT 6.S965

Taught by

MIT HAN Lab

Reviews

Start your review of Distributed Training and Gradient Compression - Part I - Lecture 13

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.