Intra-class progressive and adaptive self-distillation.

Journal: Neural networks : the official journal of the International Neural Network Society
Published Date:

Abstract

In recent years, knowledge distillation (KD) has become widely used in compressing models, training compact and efficient students to reduce computational load and training time due to the increasing parameters in deep neural networks. To minimize training costs, self-distillation has been proposed, with methods like offline-KD and online-KD requiring pre-trained teachers and multiple networks. However, these self-distillation methods often overlook feature knowledge and category information. In this paper, we introduce Intra-class Progressive and Adaptive Self-Distillation (IPASD), which transfers knowledge from the front to the back in adjacent epochs. This method extracts class-typical features and promotes compactness within classes. By integrating feature-level and logits-level knowledge into strong teacher knowledge and using ground-truth labels as supervision signals, we adaptively optimize the model. We evaluated IPASD on CIFAR-10, CIFAR-100, Tiny ImageNet, Plant Village datasets, and ImageNet showing its superiority over state-of-the-art self-distillation methods in knowledge transfer and model compression. Our codes are available at: https://github.com/JLinye/IPASD.

Authors

  • Jianping Gou
    School of Computer Science and Communication Engineering and Jiangsu Key Laboratory of Security Tech. for Industrial Cyberspace, Jiangsu University, Zhenjiang, Jiangsu, 212013, China. Electronic address: goujianping@ujs.edu.cn.
  • Jiaye Lin
    School of Computer Science and Communication Engineering, Jiangsu University, ZhenJiang, 212013, Jiangsu, China.
  • Lin Li
    Department of Medicine III, LMU University Hospital, LMU Munich, Munich, Germany.
  • Weihua Ou
    School of Big Data and Computer Science, Guizhou Normal University, Guiyang, Guizhou, 550025, China.
  • Baosheng Yu
    School of Computer Science, University of Sydney, Sydney, Australia. Electronic address: baosheng.yu@sydney.edu.au.
  • Zhang Yi