Neural learning rules for generating flexible predictions and computing the successor representation.

Journal: eLife
PMID:

Abstract

The predictive nature of the hippocampus is thought to be useful for memory-guided cognitive behaviors. Inspired by the reinforcement learning literature, this notion has been formalized as a predictive map called the successor representation (SR). The SR captures a number of observations about hippocampal activity. However, the algorithm does not provide a neural mechanism for how such representations arise. Here, we show the dynamics of a recurrent neural network naturally calculate the SR when the synaptic weights match the transition probability matrix. Interestingly, the predictive horizon can be flexibly modulated simply by changing the network gain. We derive simple, biologically plausible learning rules to learn the SR in a recurrent network. We test our model with realistic inputs and match hippocampal data recorded during random foraging. Taken together, our results suggest that the SR is more accessible in neural circuits than previously thought and can support a broad range of cognitive functions.

Authors

  • Ching Fang
    Zuckerman Institute, Department of Neuroscience, Columbia University, New York, United States.
  • Dmitriy Aronov
    Department of Neuroscience, Zuckerman Mind Brain Behavior Institute, Columbia University, New York, NY, USA.
  • L F Abbott
    Department of Neuroscience, Columbia University College of Physicians and Surgeons, New York, New York, USA.
  • Emily L Mackevicius
    Zuckerman Institute, Department of Neuroscience, Columbia University, New York, United States.