The emergence of a concept in shallow neural networks.

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

We consider restricted Boltzmann machine (RBMs) trained over an unstructured dataset made of blurred copies of definite but unavailable "archetypes" and we show that there exists a critical sample size beyond which the RBM can learn archetypes, namely the machine can successfully play as a generative model or as a classifier, according to the operational routine. In general, assessing a critical sample size (possibly in relation to the quality of the dataset) is still an open problem in machine learning. Here, restricting to the random theory, where shallow networks suffice and the "grandmother-cell" scenario is correct, we leverage the formal equivalence between RBMs and Hopfield networks, to obtain a phase diagram for both the neural architectures which highlights regions, in the space of the control parameters (i.e., number of archetypes, number of neurons, size and quality of the training set), where learning can be accomplished. Our investigations are led by analytical methods based on the statistical-mechanics of disordered systems and results are further corroborated by extensive Monte Carlo simulations.

Authors

  • Elena Agliari
    Dipartimento di Fisica, Sapienza Università di Roma, P.le A. Moro 2, 00185, Roma, Italy.
  • Francesco Alemanno
    Dipartimento di Matematica e Fisica Ennio De Giorgi, Università del Salento, Italy; C.N.R. Nanotec Lecce, Italy.
  • Adriano Barra
    Dipartimento di Fisica, Sapienza Università di Roma, P.le A. Moro 2, 00185, Roma, Italy.
  • Giordano De Marzo
    Dipartimento di Fisica, Sapienza Università di Roma, P.le A. Moro 5, 00185, Rome, Italy; Centro Ricerche Enrico Fermi, Via Panisperna 89a, 00184 Rome, Italy.