APS Logo

Modern Hopfield Networks in AI and Neurobiology

ORAL · Invited

Abstract

Modern Hopfield Networks or Dense Associative Memories are recurrent neural networks with fixed point attractor states that are described by an energy function. In contrast to conventional Hopfield Networks, their modern versions have a very large memory storage capacity, which makes them appealing tools for many problems in machine learning and cognitive and neuro-sciences. In this talk I will introduce an intuition and a mathematical formulation of this class of models, and will give examples of problems in AI that can be tackled using these new ideas. I will also explain how different individual models of this class (e.g. hierarchical memories, attention mechanism in transformers, etc.) arise from their general mathematical formulation with the Lagrangian functions.

Publication: 1. Krotov, D. and Hopfield, J.J., 2016. Dense associative memory for pattern recognition. Advances in neural information processing systems, 29.<br>2. Krotov, D. and Hopfield, J.J., 2020, September. Large Associative Memory Problem in Neurobiology and Machine Learning. In International Conference on Learning Representations.<br>3. Krotov, D., 2021. Hierarchical associative memory. arXiv preprint arXiv:2107.06446.

Presenters

  • Dmitry Krotov

    IBM Research

Authors

  • Dmitry Krotov

    IBM Research