Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore a detailed explanation of DeepMind's PonderNet, a machine learning research paper that introduces an algorithm for adaptive computation based on problem complexity. Learn about the novel approach to dynamically allocate computational steps for input samples using a recurrent architecture and trainable halting probability function. Dive into the probabilistic formulation, training methods, loss function, and experimental results. Understand how PonderNet improves performance over previous adaptive computation methods and succeeds in extrapolation tests where traditional neural networks fail. Gain insights into its applications in question-answering tasks and its potential impact on the field of machine learning and artificial intelligence.
Syllabus
- Intro & Overview
- Problem Statement
- Probabilistic formulation of dynamic halting
- Training via unrolling
- Loss function and regularization of the halting distribution
- Experimental Results
- Sensitivity to hyperparameter choice
- Discussion, Conclusion, Broader Impact
Taught by
Yannic Kilcher