Learning in the machine: To share or not to share?

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

Weight-sharing is one of the pillars behind Convolutional Neural Networks and their successes. However, in physical neural systems such as the brain, weight-sharing is implausible. This discrepancy raises the fundamental question of whether weight-sharing is necessary. If so, to which degree of precision? If not, what are the alternatives? The goal of this study is to investigate these questions, primarily through simulations where the weight-sharing assumption is relaxed. Taking inspiration from neural circuitry, we explore the use of Free Convolutional Networks and neurons with variable connection patterns. Using Free Convolutional Networks, we show that while weight-sharing is a pragmatic optimization approach, it is not a necessity in computer vision applications. Furthermore, Free Convolutional Networks match the performance observed in standard architectures when trained using properly translated data (akin to video). Under the assumption of translationally augmented data, Free Convolutional Networks learn translationally invariant representations that yield an approximate form of weight-sharing.

Authors

  • Jordan Ott
    Fowler School of Engineering, Chapman University, United States of America; Department of Computer Science, Bren School of Information and Computer Sciences, University of California, Irvine, United States of America. Electronic address: jott1@uci.edu.
  • Erik Linstead
    Schmid College of Science and Technology, Chapman University, Orange, CA, USA.
  • Nicholas LaHaye
    Fowler School of Engineering, Chapman University, United States of America. Electronic address: lahay100@mail.chapman.edu.
  • Pierre Baldi
    Department of Computer Science, Department of Biological Chemistry, University of California-Irvine, Irvine, CA 92697, USA.