Overview
Explore the AI Model Efficiency Toolkit (AIMET) in this guest lecture from Qualcomm AI Research. Dive into the world of efficient machine learning techniques for deploying neural networks on resource-constrained devices. Learn about model compression, pruning, quantization, neural architecture search, and distillation. Discover efficient training methods like gradient compression and on-device transfer learning. Examine application-specific model optimization for videos, point cloud, and NLP. Gain insights into efficient quantum machine learning. Understand the importance of AI model efficiency in the context of mobile and IoT devices. Explore community research, core technologies, and industry hardware. Get an overview of the Qualcomm AI Engine and Snap Driver. Witness demos on video understanding and super resolution. This lecture is part of the MIT 6.S965 course on TinyML and Efficient Deep Learning Computing, instructed by Song Han.
Syllabus
Introduction
Welcome
Why AI Model Efficiency
Community Research
Core Technologies
Area of Interest
Add Around
Autopart
Training
Results
Use Cases
Industry Hardware
GitHub
Training Pipeline
Whitepaper
Qualcomm AI Engine
Qualcomm Snap Driver
Demos
Video Understanding Demo
Full System View
Super Resolution
Taught by
MIT HAN Lab