Regularization of Big Neural Networks

Regularization of Big Neural Networks

UCF CRCV via YouTube Direct link

Effect of Sparsity

22 of 32

22 of 32

Effect of Sparsity

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Regularization of Big Neural Networks

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Intro
  2. 2 Big Neural Nets
  3. 3 Big Models Over-Fitting
  4. 4 Training with DropOut
  5. 5 DropOut/Connect Intuition
  6. 6 Theoretical Analysis of DropConnect
  7. 7 MNIST Results
  8. 8 Varying Size of Network
  9. 9 Varying Fraction Dropped
  10. 10 Comparison of Convergence Rates
  11. 11 Limitations of DropOut/Connect
  12. 12 Stochastic Pooling
  13. 13 Methods for Test Time
  14. 14 Varying Size of Training Set
  15. 15 Convergence / Over-Fitting
  16. 16 Street View House Numbers
  17. 17 Deconvolutional Networks
  18. 18 Recap: Sparse Coding (Patch-based)
  19. 19 Reversible Max Pooling
  20. 20 Single Layer Cost Function
  21. 21 Single Layer Inference
  22. 22 Effect of Sparsity
  23. 23 Effect of Pooling Variables
  24. 24 Talk Overview
  25. 25 Stacking the Layers
  26. 26 Two Layer Example
  27. 27 Link to Parts and Structure Models
  28. 28 Caltech 101 Experiments
  29. 29 Layer 2 Filters
  30. 30 Classification Results: Caltech 101
  31. 31 Deconvolutional + Convolutional
  32. 32 Summary

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.