Continuous transfer of neural network representational similarity for incremental learning

被引:43
|
作者
Tian, Songsong [1 ,2 ]
Li, Weijun [1 ,3 ,4 ]
Ning, Xin [1 ,3 ,4 ,5 ]
Ran, Hang [1 ]
Qin, Hong [1 ,3 ,4 ]
Tiwari, Prayag [6 ]
机构
[1] Chinese Acad Sci, Inst Semicond, Beijing 100083, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100049, Peoples R China
[3] Univ Chinese Acad Sci, Ctr Mat Sci & Optoelect Engn, Beijing 100049, Peoples R China
[4] Univ Chinese Acad Sci, Sch Integrated Circuits, Beijing 100049, Peoples R China
[5] Zhongke Ruitu Technol Co Ltd, Beijing 100096, Peoples R China
[6] Halmstad Univ, Sch Informat Technol, S-30118 Halmstad, Sweden
关键词
Incremental learning; Pre-trained model; Knowledge distillation; Neural network representation;
D O I
10.1016/j.neucom.2023.126300
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The incremental learning paradigm in machine learning has consistently been a focus of academic research. It is similar to the way in which biological systems learn, and reduces energy consumption by avoiding excessive retraining. Existing studies utilize the powerful feature extraction capabilities of pre-trained models to address incremental learning, but there remains a problem of insufficient utiliza-tion of neural network feature knowledge. To address this issue, this paper proposes a novel method called Pre-trained Model Knowledge Distillation (PMKD) which combines knowledge distillation of neu-ral network representations and replay. This paper designs a loss function based on centered kernel align-ment to transfer neural network representations knowledge from the pre-trained model to the incremental model layer-by-layer. Additionally, the use of memory buffer for Dark Experience Replay helps the model retain past knowledge better. Experiments show that PMKD achieved superior perfor-mance on various datasets and different buffer sizes. Compared to other methods, our class incremental learning accuracy reached the best performance. The open-source code is published athttps://github.-com/TianSongS/PMKD-IL.(c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Multi layered neural network with structural lateral inhibition for incremental learning and conceptualization
    Uragami, Daisuke
    Ohta, Hiroyuki
    BIOSYSTEMS, 2014, 118 : 8 - 16
  • [22] Recurrent neural network architecture with pre-synaptic inhibition for incremental learning
    Ohta, Hiroyuki
    Gunji, Yukio Pegio
    NEURAL NETWORKS, 2006, 19 (08) : 1106 - 1119
  • [23] Tree-CNN: A hierarchical Deep Convolutional Neural Network for incremental learning
    Roy, Deboleena
    Panda, Priyadarshini
    Roy, Kaushik
    NEURAL NETWORKS, 2020, 121 : 148 - 160
  • [24] INCREMENTAL LEARNING AND GENERALIZATION ABILITY OF ARTIFICIAL NEURAL NETWORK TRAINED BY FAHLMAN AND LEBIERE LEARNING ALGORITHM
    HAMAMOTO, M
    KAMRUZZAMAN, J
    KUMAGAI, Y
    HIKITA, H
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 1993, E76A (02) : 242 - 247
  • [25] IncreGNN: Incremental Graph Neural Network Learning by Considering Node and Parameter Importance
    Wei, Di
    Gu, Yu
    Song, Yumeng
    Song, Zhen
    Li, Fangfang
    Yu, Ge
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2022, PT I, 2022, : 739 - 746
  • [26] An Imbalance Modified Convolutional Neural Network With Incremental Learning for Chemical Fault Diagnosis
    Gu, Xiaohua
    Zhao, Yanli
    Yang, Guang
    Li, Lusi
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2022, 18 (06) : 3630 - 3639
  • [27] Incremental Learning in Deep Convolutional Neural Networks Using Partial Network Sharing
    Sarwar, Syed Shakib
    Ankit, Aayush
    Roy, Kaushik
    IEEE ACCESS, 2020, 8 (08): : 4615 - 4628
  • [28] An incremental-learning neural network for the classification of remote-sensing images
    Bruzzone, L
    Prieto, DF
    PATTERN RECOGNITION LETTERS, 1999, 20 (11-13) : 1241 - 1248
  • [29] Gaussian ARTMAP: A neural network for past incremental learning of noisy multidimensional maps
    Williamson, JR
    NEURAL NETWORKS, 1996, 9 (05) : 881 - 897
  • [30] Incremental constructive ridgelet neural network
    Yang, Shuyuan
    Wang, Min
    Jiao, Licheng
    NEUROCOMPUTING, 2008, 72 (1-3) : 367 - 377