Overview
Syllabus
Intro
What this talk is about The problem: • Many GPUs available for deep neural network (DNN) training . Each has a different cost and performance
A Cambrian explosion in hardware for training
Choosing a GPU: The paradox of choice
Key observations • Deep learning users may already have an existing GPU
Habitat: A runtime-based performance predictor
One last wrinkle: Kernel-varying operations Wave scaling assumes the same kernel is used across GPUS
Evaluation
How accurate is Habitat?
Rent a GPU in the cloud? Scenario: Want to train GNMT, have access to a P4000. Which cloud GPU to use, if any?
Key takeaways . DNN computation is special (repetitive), enabling new analysis opportunities
Taught by
USENIX