Modern Hopfield Networks in AI and Neurobiology
ORAL · Invited
Abstract
Modern Hopfield Networks or Dense Associative Memories are recurrent neural networks with fixed point attractor states that are described by an energy function. In contrast to conventional Hopfield Networks, their modern versions have a very large memory storage capacity, which makes them appealing tools for many problems in machine learning and cognitive and neuro-sciences. In this talk I will introduce an intuition and a mathematical formulation of this class of models, and will give examples of problems in AI that can be tackled using these new ideas. I will also explain how different individual models of this class (e.g. hierarchical memories, attention mechanism in transformers, etc.) arise from their general mathematical formulation with the Lagrangian functions.
–
Publication: 1. Krotov, D. and Hopfield, J.J., 2016. Dense associative memory for pattern recognition. Advances in neural information processing systems, 29.<br>2. Krotov, D. and Hopfield, J.J., 2020, September. Large Associative Memory Problem in Neurobiology and Machine Learning. In International Conference on Learning Representations.<br>3. Krotov, D., 2021. Hierarchical associative memory. arXiv preprint arXiv:2107.06446.
Presenters
-
Dmitry Krotov
IBM Research
Authors
-
Dmitry Krotov
IBM Research