Skip to main content
Video s3
ID 2334

Modern Hopfield Networks

    Details
    Presenter(s)
    Sepp Hochreiter Headshot
    Display Name
    Sepp Hochreiter
    Affiliation
    Affiliation
    Johannes Kepler University
    Country
    Abstract

    We propose a new paradigm for deep learning by equipping each layer of a deep learning architecture with modern Hopfield networks. The new paradigm is a new powerful concept comprising functionalities like pooling, memory, and attention for each layer. Associative memories date back to the 1960/70s and became popular through Hopfield Networks in 1982. Recently, we saw a renaissance of Hopfield Networks, the modern Hopfield Networks, with a tremendously increased storage capacity and an extremely fast convergence. We generalize modern Hopfield Networks with exponential storage capacity to continuous patterns. Their update rule ensures global convergence to local energy minima and they converge in one update step with exponentially low error. Surprisingly, the transformer attention mechanism is equal to the update rule of our new modern Hopfield Network with continuous states. The new modern Hopfield network can be integrated into deep learning architectures as layers to allow the storage of and access to raw input data, intermediate results, or learned prototypes.

    Description

    Please find additional materials regarding this topic below:

    github: https://github.com/ml-jku/hopfield-layers

    arXiv: “Hopfield Networks is All You Need”, https://arxiv.org/abs/2008.02217

    blog: https://ml-jku.github.io/hopfield-layers/

    Videos: “Modern Hopfield Networks” (S. Hochreiter)

    https://youtu.be/bsdPZJKOlQs

    “Hopfield Networks in 2021” (S. Hochreiter & D. Krotov)

    https://youtu.be/k3YmWrK6wxo

    “Performers & Memory” (S. Hochreiter & K. Choromanski)

    https://youtu.be/BuDhitKZ_YY