The Physics of Associative Memory - Exploring Hopfield Networks
Offered By: Artem Kirsanov via YouTube
Course Description
Overview
Explore the concept of Hopfield networks in this 27-minute video lecture on the physics of associative memory. Delve into a foundational model underlying key ideas in neuroscience and machine learning, including Boltzmann machines and Dense associative memory. Begin with an introduction to the protein folding paradox, then progress through energy definition, Hopfield network architecture, inference, and learning processes. Examine the limitations and perspectives of this model before concluding with a brief discussion on related topics. Gain insights into the intersection of physics, neuroscience, and machine learning through this comprehensive exploration of associative memory systems.
Syllabus
Introduction
Protein folding paradox
Energy definition
Hopfield network architecture
Inference
Learning
Limitations & Perspective
Shortform
Outro
Taught by
Artem Kirsanov
Related Courses
Introduction to Artificial IntelligenceStanford University via Udacity Natural Language Processing
Columbia University via Coursera Probabilistic Graphical Models 1: Representation
Stanford University via Coursera Computer Vision: The Fundamentals
University of California, Berkeley via Coursera Learning from Data (Introductory Machine Learning course)
California Institute of Technology via Independent