Poster
Nonparametric Modern Hopfield Models
Jerry Yao-Chieh Hu · Bo-Yu Chen · Dennis Wu · Feng Ruan · Han Liu
East Exhibition Hall A-B #E-3403
Modern AI systems often depend on an “attention” step that looks at every piece of data all at once, a procedure that quickly becomes slow and costly as models grow. Our work recasts a powerful alternative, the modern Hopfield network, as a simple form of pattern-matching regression. This fresh view lets us build sparse-structured Hopfield memories that ignore the many data items that do not matter, slashing the usual quadratic compute cost to nearly linear while keeping three key benefits: (i) one-shot recall of stored patterns, (ii) strong resistance to noise, and (iii) the ability to hold an exponential number of memories. We back these claims with new mathematical error bounds and with experiments on images, time-series data and multiple-instance learning tasks. In practice, our layers can drop into today’s deep-learning code as a cheaper replacement for attention, delivering similar or better accuracy while using far less computation.