Continual learning with Bayesian compression for shared and private latent representations.

Journal: Neural networks : the official journal of the International Neural Network Society
PMID:

Abstract

This paper proposes a new continual learning method with Bayesian Compression for Shared and Private Latent Representations (BCSPLR), which learns a compact model structure while preserving the accuracy. In Shared and Private Latent Representations (SPLR), task-invariant and task-specific latent representations are efficiently learned to avoid catastrophic forgetting, whereas SPLR produces point estimates of parameters and manually tunes multiple hyperparameters. To overcome these problems, a principle framework is used to develop Bayesian Compression for SPLR, which can learn task-specific latent features with significant changes and task-invariant latent representations with small changes in the continual learning scenarios. To evaluate our algorithm, MNIST, CIFAR100 and ImageNet100 datasets are used to verify our BCSPLR. From these experiments, our model can learn shared and private compact structure which means fewer parameters, and obtain comparable training time, importantly, the model performance is also excellent compared with the state-of-the-art continual learning algorithms.

Authors

  • Yang Yang
    Department of Gastrointestinal Surgery, The Third Hospital of Hebei Medical University, Shijiazhuang, China.
  • Dandan Guo
    Department of Chemistry, Xixi Campus, Zhejiang University, Hangzhou, 310028, China.
  • Bo Chen
  • Dexiu Hu
    PLA Strategic Support Force Information Engineering University, Zhengzhou, Henan, 450001, China. Electronic address: paper_hdx@126.com.