Visual prototypes in the ventral stream are attuned to complexity and gaze behavior.

Journal: Nature communications
PMID:

Abstract

Early theories of efficient coding suggested the visual system could compress the world by learning to represent features where information was concentrated, such as contours. This view was validated by the discovery that neurons in posterior visual cortex respond to edges and curvature. Still, it remains unclear what other information-rich features are encoded by neurons in more anterior cortical regions (e.g., inferotemporal cortex). Here, we use a generative deep neural network to synthesize images guided by neuronal responses from across the visuocortical hierarchy, using floating microelectrode arrays in areas V1, V4 and inferotemporal cortex of two macaque monkeys. We hypothesize these images ("prototypes") represent such predicted information-rich features. Prototypes vary across areas, show moderate complexity, and resemble salient visual attributes and semantic content of natural images, as indicated by the animals' gaze behavior. This suggests the code for object recognition represents compressed features of behavioral relevance, an underexplored aspect of efficient coding.

Authors

  • Olivia Rose
    Department of Neuroscience, Washington University School of Medicine, St. Louis, MO, USA.
  • James Johnson
    Department of Neuroscience, Washington University School of Medicine, St. Louis, MO, USA.
  • Binxu Wang
    Center for Bioinformatics, National Laboratory of Protein Engineering and Plant Genetic Engineering, School of Life Sciences, Peking University, Beijing, 100871, China, and Neuroscience Program, Washington University, St. Louis, MO 63130, U.S.A. lio50328@126.com.
  • Carlos R Ponce
    Department of Neuroscience, Washington University School of Medicine, St. Louis, MO, USA. carlos_ponce@hms.harvard.edu.