Continuous transfer of neural network representational similarity for incremental learning

被引:43
|
作者
Tian, Songsong [1 ,2 ]
Li, Weijun [1 ,3 ,4 ]
Ning, Xin [1 ,3 ,4 ,5 ]
Ran, Hang [1 ]
Qin, Hong [1 ,3 ,4 ]
Tiwari, Prayag [6 ]
机构
[1] Chinese Acad Sci, Inst Semicond, Beijing 100083, Peoples R China
[2] Univ Chinese Acad Sci, Sch Elect Elect & Commun Engn, Beijing 100049, Peoples R China
[3] Univ Chinese Acad Sci, Ctr Mat Sci & Optoelect Engn, Beijing 100049, Peoples R China
[4] Univ Chinese Acad Sci, Sch Integrated Circuits, Beijing 100049, Peoples R China
[5] Zhongke Ruitu Technol Co Ltd, Beijing 100096, Peoples R China
[6] Halmstad Univ, Sch Informat Technol, S-30118 Halmstad, Sweden
关键词
Incremental learning; Pre-trained model; Knowledge distillation; Neural network representation;
D O I
10.1016/j.neucom.2023.126300
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The incremental learning paradigm in machine learning has consistently been a focus of academic research. It is similar to the way in which biological systems learn, and reduces energy consumption by avoiding excessive retraining. Existing studies utilize the powerful feature extraction capabilities of pre-trained models to address incremental learning, but there remains a problem of insufficient utiliza-tion of neural network feature knowledge. To address this issue, this paper proposes a novel method called Pre-trained Model Knowledge Distillation (PMKD) which combines knowledge distillation of neu-ral network representations and replay. This paper designs a loss function based on centered kernel align-ment to transfer neural network representations knowledge from the pre-trained model to the incremental model layer-by-layer. Additionally, the use of memory buffer for Dark Experience Replay helps the model retain past knowledge better. Experiments show that PMKD achieved superior perfor-mance on various datasets and different buffer sizes. Compared to other methods, our class incremental learning accuracy reached the best performance. The open-source code is published athttps://github.-com/TianSongS/PMKD-IL.(c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Incremental clustering algorithm of neural network
    Liu P.
    Tang J.
    Xie S.
    Wang T.
    Guofang Keji Daxue Xuebao, 5 (137-142): : 137 - 142
  • [32] An incremental neural network with a reduced architecture
    Ciarelli, Patrick Marques
    Oliveira, Elias
    Salles, Evandro O. T.
    NEURAL NETWORKS, 2012, 35 : 70 - 81
  • [33] Incremental Similarity for real-time on-line incremental learning systems
    Reznakova, Marta
    Tencer, Lukas
    Cheriet, Mohamed
    PATTERN RECOGNITION LETTERS, 2016, 74 : 61 - 67
  • [34] Incremental Learning Based on Dual-Branch Network
    Dong, Mingda
    Zhang, Zhizhong
    Xie, Yuan
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT III, 2024, 14427 : 263 - 272
  • [35] INCREMENTAL LEARNING OF CONVOLUTIONAL NEURAL NETWORKS
    Medera, Dusan
    Babinec, Stefan
    IJCCI 2009: PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON COMPUTATIONAL INTELLIGENCE, 2009, : 547 - +
  • [36] A Self Adaptive Incremental learning Fuzzy Neural Network Based on the Influence of a Fuzzy Rule
    Rong, Hu
    Xiang, Xu
    Ye, Xia
    2015 INTERNATIONAL CONFERENCE ON INTELLIGENT INFORMATION HIDING AND MULTIMEDIA SIGNAL PROCESSING (IIH-MSP), 2015, : 354 - 359
  • [37] An Imbalance Modified Deep Neural Network With Dynamical Incremental Learning for Chemical Fault Diagnosis
    Hu, Zhixin
    Jiang, Peng
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2019, 66 (01) : 540 - 550
  • [38] Self-Improving Generative Artificial Neural Network for Pseudorehearsal Incremental Class Learning
    Mellado, Diego
    Saavedra, Carolina
    Chabert, Steren
    Torres, Romina
    Salas, Rodrigo
    ALGORITHMS, 2019, 12 (10)
  • [39] Continuous Recognition with Incremental Learning on Grassmann Manifolds
    Azary, Sherif
    Savakis, Andreas
    2017 IEEE 60TH INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS (MWSCAS), 2017, : 1477 - 1480
  • [40] A modular type network for incremental learning
    Ishihara, S
    Nagano, T
    ICONIP'98: THE FIFTH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING JOINTLY WITH JNNS'98: THE 1998 ANNUAL CONFERENCE OF THE JAPANESE NEURAL NETWORK SOCIETY - PROCEEDINGS, VOLS 1-3, 1998, : 1651 - 1654