Disturbance-immune weight sharing for neural architecture search.

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

Neural architecture search (NAS) has gained increasing attention in the community of architecture design. One of the key factors behind the success lies in the training efficiency brought by the weight sharing (WS) technique. However, WS-based NAS methods often suffer from a performance disturbance (PD) issue. That is, the training of subsequent architectures inevitably disturbs the performance of previously trained architectures due to the partially shared weights. This leads to inaccurate performance estimation for the previous architectures, which makes it hard to learn a good search strategy. To alleviate the performance disturbance issue, we propose a new disturbance-immune update strategy for model updating. Specifically, to preserve the knowledge learned by previous architectures, we constrain the training of subsequent architectures in an orthogonal space via orthogonal gradient descent. Equipped with this strategy, we propose a novel disturbance-immune training scheme for NAS. We theoretically analyze the effectiveness of our strategy in alleviating the PD risk. Extensive experiments on CIFAR-10 and ImageNet verify the superiority of our method.

Authors

  • Shuaicheng Niu
    South China University of Technology, China; Pazhou Laboratory, China; Key Laboratory of Big Data and Intelligent Robot, Ministry of Education, China. Electronic address: sensc@mail.scut.edu.cn.
  • Jiaxiang Wu
    Tencent AI Lab, Shenzhen, China.
  • Yifan Zhang
    Department of Food Science and Nutrition, Zhejiang Key Laboratory for Agro-Food Processing, Zhejiang University, Hangzhou, Zhejiang 310058, China.
  • Yong Guo
    Department of Urology, The First Hospital of Shijiazhuang, Shijiazhuang 050011, China.
  • Peilin Zhao
    Tencent AI Lab, China. Electronic address: masonzhao@tencent.com.
  • Junzhou Huang
  • Mingkui Tan