Using deep reinforcement learning to reveal how the brain encodes abstract state-space representations in high-dimensional environments.

Journal: Neuron
PMID:

Abstract

Humans possess an exceptional aptitude to efficiently make decisions from high-dimensional sensory observations. However, it is unknown how the brain compactly represents the current state of the environment to guide this process. The deep Q-network (DQN) achieves this by capturing highly nonlinear mappings from multivariate inputs to the values of potential actions. We deployed DQN as a model of brain activity and behavior in participants playing three Atari video games during fMRI. Hidden layers of DQN exhibited a striking resemblance to voxel activity in a distributed sensorimotor network, extending throughout the dorsal visual pathway into posterior parietal cortex. Neural state-space representations emerged from nonlinear transformations of the pixel space bridging perception to action and reward. These transformations reshape axes to reflect relevant high-level features and strip away information about task-irrelevant sensory features. Our findings shed light on the neural encoding of task representations for decision-making in real-world situations.

Authors

  • Logan Cross
    Computation and Neural Systems, California Institute of Technology, Pasadena, CA 91125, USA. Electronic address: lcross@caltech.edu.
  • Jeff Cockburn
    Division of Humanities and Social Sciences, California Institute of Technology, Pasadena, CA 91125, USA.
  • Yisong Yue
    Department of Computing and Mathematical Sciences, California Institute of Technology, Pasadena, CA 91125, USA.
  • John P O'Doherty
    Division of Humanities and Social Sciences, California Institute of Technology, Pasadena, CA 91125, USA.