Adaptive lightweight network construction method for Self-Knowledge Distillation

被引:0
|
作者
Lu, Siyuan [1 ]
Zeng, Weiliang [1 ]
Li, Xueshi [1 ]
Ou, Jiajun [1 ]
机构
[1] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Guangdong, Peoples R China
关键词
Deep learning; Knowledge Distillation; Neural network architecture design;
D O I
10.1016/j.neucom.2025.129477
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-Knowledge Distillation (self-KD) has become a promising method for neural network compression due to its advances in computational efficiency. Nevertheless, its applicability is constrained by the inherent inflexibility of the network architecture and the absence of quantitative metrics to evaluate the distillability of the architecture. To address these problems, a two-stage adaptive dynamic distillation network framework (ADDN) is proposed to adapt the architecture based on the distillability of the current architecture, containing a hypernetwork topology construction stage and a subnetwork training stage. To evaluate the distillability of current architectures without necessitating extensive training, we propose a set of low-cost distillability metrics that evaluate architectures from the perspective of architectural similarity and clustering ability. Furthermore, to simplify the hypernetwork structure and reduce the complexity of the construction process, a hierarchical filtering module is introduced to dynamically refine and remove candidate operations within the architecture incrementally, contingent upon the distillability of the current architecture. To validate the effectiveness of our approach, we conduct extensive experiments on various image classification datasets and compare with current works. Experimental results demonstrate that the self-knowledge distillation network architecture obtained by our proposed methodology simultaneously attains superior distillability and efficiency while significantly curtailing construction expenses.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] A Lightweight Approach for Network Intrusion Detection based on Self-Knowledge Distillation
    Yang, Shuo
    Zheng, Xinran
    Xu, Zhengzhuo
    Wang, Xingjun
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 3000 - 3005
  • [2] A Lightweight Convolution Network with Self-Knowledge Distillation for Hyperspectral Image Classification
    Xu, Hao
    Cao, Guo
    Deng, Lindiao
    Ding, Lanwei
    Xu, Ling
    Pan, Qikun
    Shang, Yanfeng
    FOURTEENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING, ICGIP 2022, 2022, 12705
  • [3] Many-objective evolutionary self-knowledge distillation with adaptive branch fusion method
    Bai, Jiayuan
    Zhang, Yi
    INFORMATION SCIENCES, 2024, 669
  • [4] Neighbor self-knowledge distillation
    Liang, Peng
    Zhang, Weiwei
    Wang, Junhuang
    Guo, Yufeng
    INFORMATION SCIENCES, 2024, 654
  • [5] Lightweight Human Pose Estimation Based on Densely Guided Self-Knowledge Distillation
    Wu, Mingyue
    Zhao, Zhong-Qiu
    Li, Jiajun
    Tian, Weidong
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT II, 2023, 14255 : 421 - 433
  • [6] Self-knowledge distillation with dimensional history knowledge
    Wenke Huang
    Mang Ye
    Zekun Shi
    He Li
    Bo Du
    Science China Information Sciences, 2025, 68 (9)
  • [7] Uncertainty Driven Adaptive Self-Knowledge Distillation for Medical Image Segmentation
    Guo, Xutao
    Wang, Mengqi
    Xiang, Yang
    Yang, Yanwu
    Ye, Chenfei
    Wang, Haijun
    Ma, Ting
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2025,
  • [8] Self-knowledge distillation via dropout
    Lee, Hyoje
    Park, Yeachan
    Seo, Hyun
    Kang, Myungjoo
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 233
  • [9] Dual teachers for self-knowledge distillation
    Li, Zheng
    Li, Xiang
    Yang, Lingfeng
    Song, Renjie
    Yang, Jian
    Pan, Zhigeng
    PATTERN RECOGNITION, 2024, 151
  • [10] Training a thin and shallow lane detection network with self-knowledge distillation
    Dai, Xuerui
    Yuan, Xue
    Wei, Xueye
    JOURNAL OF ELECTRONIC IMAGING, 2021, 30 (01)