Generative adversarial networks with decoder-encoder output noises.

Journal: Neural networks : the official journal of the International Neural Network Society
PMID:

Abstract

In recent years, research on image generation has been developing very fast. The generative adversarial network (GAN) emerges as a promising framework, which uses adversarial training to improve the generative ability of its generator. However, since GAN and most of its variants use randomly sampled noises as the input of their generators, they have to learn a mapping function from a whole random distribution to the image manifold. As the structures of the random distribution and the image manifold are generally different, this results in GAN and its variants difficult to train and converge. In this paper, we propose a novel deep model called generative adversarial networks with decoder-encoder output noises (DE-GANs), which take advantage of both the adversarial training and the variational Bayesian inference to improve GAN and its variants on image generation performances. DE-GANs use a pre-trained decoder-encoder architecture to map the random noise vectors to informative ones and feed them to the generator of the adversarial networks. Since the decoder-encoder architecture is trained with the same data set as the generator, its output vectors, as the inputs of the generator, could carry the intrinsic distribution information of the training images, which greatly improves the learnability of the generator and the quality of the generated images. Extensive experiments demonstrate the effectiveness of the proposed model, DE-GANs.

Authors

  • Guoqiang Zhong
    Department of Computer Science and Technology, Ocean University of China, Qingdao 266100, China. Electronic address: gqzhong@ouc.edu.cn.
  • Wei Gao
    Andrew and Peggy Cherng Department of Medical Engineering, Division of Engineering and Applied Science, California Institute of Technology, Pasadena, CA, USA.
  • Yongbin Liu
    Department of Computer Science and Technology, Ocean University of China, Qingdao 266100, China. Electronic address: liuyongbin@stu.ouc.edu.cn.
  • Youzhao Yang
    Department of Computer Science and Technology, Fudan University, Shanghai 200433, China. Electronic address: yzyang17@fudan.edu.cn.
  • Da-Han Wang
    Fujian Key Laboratory of Pattern Recognition and Image Understanding, Xiamen University of Technology, Xiamen 361024, China. Electronic address: wangdh@xmut.edu.cn.
  • Kaizhu Huang
    Xi'an Jiaotong-Liverpool University, Suzhou, Jiangsu, China. Electronic address: Kaizhu.Huang@xjtlu.edu.cn.