Explaining Neural Networks Using Attentive Knowledge Distillation

被引:4
|
作者
Lee, Hyeonseok [1 ]
Kim, Sungchan [1 ,2 ]
机构
[1] Jeonbuk Natl Univ, Div Comp Sci & Engn, Jeonju Si 54896, Jeollabuk Do, South Korea
[2] Jeonbuk Natl Univ, Res Ctr Artificial Intelligence Technol, Jeonju Si 54896, Jeollabuk Do, South Korea
基金
新加坡国家研究基金会;
关键词
deep neural networks; visual explanation; attention; knowledge distillation; fine-grained classification;
D O I
10.3390/s21041280
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Explaining the prediction of deep neural networks makes the networks more understandable and trusted, leading to their use in various mission critical tasks. Recent progress in the learning capability of networks has primarily been due to the enormous number of model parameters, so that it is usually hard to interpret their operations, as opposed to classical white-box models. For this purpose, generating saliency maps is a popular approach to identify the important input features used for the model prediction. Existing explanation methods typically only use the output of the last convolution layer of the model to generate a saliency map, lacking the information included in intermediate layers. Thus, the corresponding explanations are coarse and result in limited accuracy. Although the accuracy can be improved by iteratively developing a saliency map, this is too time-consuming and is thus impractical. To address these problems, we proposed a novel approach to explain the model prediction by developing an attentive surrogate network using the knowledge distillation. The surrogate network aims to generate a fine-grained saliency map corresponding to the model prediction using meaningful regional information presented over all network layers. Experiments demonstrated that the saliency maps are the result of spatially attentive features learned from the distillation. Thus, they are useful for fine-grained classification tasks. Moreover, the proposed method runs at the rate of 24.3 frames per second, which is much faster than the existing methods by orders of magnitude.
引用
收藏
页码:1 / 17
页数:17
相关论文
共 50 条
  • [41] Uncertainty-aware visually-attentive navigation using deep neural networks
    Nguyen, Huan
    Andersen, Rasmus
    Boukas, Evangelos
    Alexis, Kostas
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2024, 43 (06): : 840 - 872
  • [42] Closing the Neural-Symbolic Cycle: Knowledge Extraction, User Intervention and Distillation from Convolutional Neural Networks
    Ngan, Kwun Ho
    Phelan, James
    Mansouri-Benssassi, Esma
    Townsend, Joe
    Garcez, Artur d'Avila
    NEURAL-SYMBOLIC LEARNING AND REASONING 2023, NESY 2023, 2023,
  • [43] Self-Distillation: Towards Efficient and Compact Neural Networks
    Zhang, Linfeng
    Bao, Chenglong
    Ma, Kaisheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (08) : 4388 - 4403
  • [44] Dual Knowledge Distillation for neural machine translation
    Wan, Yuxian
    Zhang, Wenlin
    Li, Zhen
    Zhang, Hao
    Li, Yanxia
    COMPUTER SPEECH AND LANGUAGE, 2024, 84
  • [45] Counterclockwise block-by-block knowledge distillation for neural network compression
    Lan, Xiaowei
    Zeng, Yalin
    Wei, Xiaoxia
    Zhang, Tian
    Wang, Yiwen
    Huang, Chao
    He, Weikai
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [46] Melanoma Breslow Thickness Classification Using Ensemble-Based Knowledge Distillation With Semi-Supervised Convolutional Neural Networks
    Dominguez-Morales, Juan P.
    Hernandez-Rodriguez, Juan-Carlos
    Duran-Lopez, Lourdes
    Conejo-Mir, Julian
    Pereyra-Rodriguez, Jose-Juan
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2025, 29 (01) : 443 - 455
  • [47] Deep Convolutional Neural Networks Based on Knowledge Distillation for Offline Handwritten Chinese Character Recognition
    He, Hongli
    Zhu, Zongnan
    Li, Zhuo
    Dan, Yongping
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2024, 28 (02) : 231 - 238
  • [48] Synthetic data generation method for data-free knowledge distillation in regression neural networks
    Zhou, Tianxun
    Chiam, Keng-Hwee
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 227
  • [49] MULTI-TEACHER KNOWLEDGE DISTILLATION FOR COMPRESSED VIDEO ACTION RECOGNITION ON DEEP NEURAL NETWORKS
    Wu, Meng-Chieh
    Chiu, Ching-Te
    Wu, Kun-Hsuan
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2202 - 2206
  • [50] Attentive differential convolutional neural networks for crowd flow prediction
    Mo, Jiqian
    Gong, Zhiguo
    Chen, Junyang
    KNOWLEDGE-BASED SYSTEMS, 2022, 258