Deep Latent-Variable Kernel Learning.

Journal: IEEE transactions on cybernetics
Published Date:

Abstract

Deep kernel learning (DKL) leverages the connection between the Gaussian process (GP) and neural networks (NNs) to build an end-to-end hybrid model. It combines the capability of NN to learn rich representations under massive data and the nonparametric property of GP to achieve automatic regularization that incorporates a tradeoff between model fit and model complexity. However, the deterministic NN encoder may weaken the model regularization of the following GP part, especially on small datasets, due to the free latent representation. We, therefore, present a complete deep latent-variable kernel learning (DLVKL) model wherein the latent variables perform stochastic encoding for regularized representation. We further enhance the DLVKL from two aspects: 1) the expressive variational posterior through neural stochastic differential equation (NSDE) to improve the approximation quality and 2) the hybrid prior taking knowledge from both the SDE prior and the posterior to arrive at a flexible tradeoff. Extensive experiments imply that DLVKL-NSDE performs similar to the well-calibrated GP on small datasets, and shows superiority on large datasets.

Authors

  • Haitao Liu
    Key Disciplines Lab of Novel Micro-nano Devices and System Technology, Chongqing University, Chongqing 400030, China; Key Laboratory for Optoelectronic Technology & System of Ministry of Education, Chongqing University, Chongqing 400044, China.
  • Yew-Soon Ong
    Rolls-Royce@NTU Corporate Lab c/o, School of Computer Engineering, Nanyang Technological University, Singapore. Electronic address: ASYSOng@ntu.edu.sg.
  • Xiaomo Jiang
  • Xiaofang Wang
    Hebei University of Chinese Medicine, Shijiazhuang, China.