Correlational Neural Networks.

Journal: Neural computation
Published Date:

Abstract

Common representation learning (CRL), wherein different descriptions (or views) of the data are embedded in a common subspace, has been receiving a lot of attention recently. Two popular paradigms here are canonical correlation analysis (CCA)-based approaches and autoencoder (AE)-based approaches. CCA-based approaches learn a joint representation by maximizing correlation of the views when projected to the common subspace. AE-based methods learn a common representation by minimizing the error of reconstructing the two views. Each of these approaches has its own advantages and disadvantages. For example, while CCA-based approaches outperform AE-based approaches for the task of transfer learning, they are not as scalable as the latter. In this work, we propose an AE-based approach, correlational neural network (CorrNet), that explicitly maximizes correlation among the views when projected to the common subspace. Through a series of experiments, we demonstrate that the proposed CorrNet is better than AE and CCA with respect to its ability to learn correlated common representations. We employ CorrNet for several cross-language tasks and show that the representations learned using it perform better than the ones learned using other state-of-the-art approaches.

Authors

  • Sarath Chandar
    University of Montreal, Montreal QC H3T 1J4, Canada apsarathchandar@gmail.com.
  • Mitesh M Khapra
    IBM Research India, Bangalore 560077, India mikhapra@in.ibm.com.
  • Hugo Larochelle
    University of Sherbrooke, Sherbrooke QC J1K 2R1, Canada hugo.larochelle@usherbrooke.ca.
  • Balaraman Ravindran
    Department of Computer Science and Engineering, Indian Institute of Technology, Madras, Chennai, India.