Brain-optimized deep neural network models of human visual areas learn non-hierarchical representations.

Journal: Nature communications
Published Date:

Abstract

Deep neural networks (DNNs) optimized for visual tasks learn representations that align layer depth with the hierarchy of visual areas in the primate brain. One interpretation of this finding is that hierarchical representations are necessary to accurately predict brain activity in the primate visual system. To test this interpretation, we optimized DNNs to directly predict brain activity measured with fMRI in human visual areas V1-V4. We trained a single-branch DNN to predict activity in all four visual areas jointly, and a multi-branch DNN to predict each visual area independently. Although it was possible for the multi-branch DNN to learn hierarchical representations, only the single-branch DNN did so. This result shows that hierarchical representations are not necessary to accurately predict human brain activity in V1-V4, and that DNNs that encode brain-like visual representations may differ widely in their architecture, ranging from strict serial hierarchies to multiple independent branches.

Authors

  • Ghislain St-Yves
    Medical University of South Carolina, Charleston, SC, USA. Electronic address: stayves@musc.edu.
  • Emily J Allen
    Center for Magnetic Resonance Research(CMRR), Department of Radiology, University of Minnesota, Minneapolis, Minnesota, USA; Department of Neuroscience, University of Minnesota, Minneapolis, Minnesota, USA.
  • Yihan Wu
    Center for Magnetic Resonance Research(CMRR), Department of Radiology, University of Minnesota, Minneapolis, Minnesota, USA.
  • Kendrick Kay
    Center for Magnetic Resonance Research(CMRR), Department of Radiology, University of Minnesota, Minneapolis, Minnesota, USA.
  • Thomas Naselaris
    Medical University of South Carolina, Charleston, SC, USA. Electronic address: tnaselar@musc.edu.