Overview
Explore a comprehensive video explanation of the research paper "Deep Ensembles: A Loss Landscape Perspective" in this 47-minute analysis. Delve into the effectiveness of Deep Ensembles in improving neural network generalization, outperforming Bayesian Networks, and their unique ability to capture the non-convex loss landscape. Follow along as the video breaks down key concepts including the solution space of deep networks, the ensemble effect, and experimental setups. Examine detailed comparisons between independent solutions, baselines, and various ensembling methods. Gain insights into weight space cross-sections, the relationship between diversity and accuracy, and the significance of random initializations in exploring diverse modes in function space. Conclude with a thorough discussion of the paper's findings and their implications for deep learning research and practice.
Syllabus
- Intro & Overview
- Deep Ensembles
- The Solution Space of Deep Networks
- Bayesian Models
- The Ensemble Effect
- Experiment Setup
- Solution Equality While Training
- Tracking Multiple Trajectories
- Similarity of Independent Solutions
- Comparison to Baselines
- Weight Space Cross-Sections
- Diversity vs Accuracy
- Comparing Ensembling Methods
- Conclusion & Comments
Taught by
Yannic Kilcher