Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

AutoMoE - Neural Architecture Search for Efficient Sparsely Activated Transformers

AutoML Seminars via YouTube

Overview

Learn about neural architecture search for efficient sparsely activated transformers in this 47-minute research seminar presentation. Explore how AutoMoE, a novel framework, introduces sparse architectures with conditional computation in the NAS search space, moving beyond traditional dense architectures where all network weights are activated for every input. Discover how AutoMoE-generated sparse models achieve 4x FLOPs reduction and equivalent CPU speedups compared to manually designed Transformers while maintaining BLEU score parity on neural machine translation benchmark datasets. Delve into the heterogeneous search space combining dense and sparsely activated Transformer modules, examining crucial aspects like expert quantity, placement, and sizing for adaptive computation. Access the complete implementation, including code, data, and trained models, through the provided GitHub repository and research paper.

Syllabus

Subho Mukherjee: "AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers"

Taught by

AutoML Seminars

Reviews

Start your review of AutoMoE - Neural Architecture Search for Efficient Sparsely Activated Transformers

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.