Predictive coding with memory-based targets

PI: Blake Richards
Co-PI: Richard Zemel

Abstract

 This project extends predictive coding—a biologically plausible theory of cortical computation—from static data to sequential data by incorporating an associative memory module. The model is trained in an autoregressive and self-supervised manner, allowing it to learn rich representations of sequential data and recall relevant information from past inputs. Importantly, the approach does not rely on backpropagation through time, making it more biologically plausible and memory efficient as sequences grow longer. The architecture uses a DeltaNet-style associative memory that stores and retrieves key-value pairs, with a learned modulation of plasticity analogous to neuromodulation.

Picture1

Figure 1: One of our proposed architectures, where M represents the memory module that stores long-context information.

Publications

In progress

Resources

In progress