Overview
Save Big on Coursera Plus. 7,000+ courses at $160 off. Limited Time Only!
Explore the concept of Hopfield networks in this 27-minute video lecture on the physics of associative memory. Delve into a foundational model underlying key ideas in neuroscience and machine learning, including Boltzmann machines and Dense associative memory. Begin with an introduction to the protein folding paradox, then progress through energy definition, Hopfield network architecture, inference, and learning processes. Examine the limitations and perspectives of this model before concluding with a brief discussion on related topics. Gain insights into the intersection of physics, neuroscience, and machine learning through this comprehensive exploration of associative memory systems.
Syllabus
Introduction
Protein folding paradox
Energy definition
Hopfield network architecture
Inference
Learning
Limitations & Perspective
Shortform
Outro
Taught by
Artem Kirsanov