A class-incremental learning approach for learning feature-compatible embeddings

被引:6
作者
An, Hongchao [1 ]
Yang, Jing [1 ,2 ]
Zhang, Xiuhua [1 ]
Ruan, Xiaoli [2 ]
Wu, Yuankai [3 ]
Li, Shaobo [1 ,2 ]
Hu, Jianjun [4 ]
机构
[1] Guizhou Univ, Sch Mech Engn, Guiyang 550025, Guizhou, Peoples R China
[2] Guizhou Univ, State Key Lab Publ Big Data, Minist Educ, Guiyang 550025, Guizhou, Peoples R China
[3] Sichuan Univ, Natl Key Lab Fundamental Sci Synthet Vis, Chengdu 610065, Sichuan, Peoples R China
[4] Univ South Carolina, Dept Comp Sci & Engn, Columbia, SC 29208 USA
基金
中国国家自然科学基金;
关键词
Class-incremental learning; Catastrophic forgetting; Feature embedding; Knowledge distillation;
D O I
10.1016/j.neunet.2024.106685
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans have the ability to constantly learn new knowledge. However, for artificial intelligence, trying to continuously learn new knowledge usually results in catastrophic forgetting, the existing regularization- based and dynamic structure-based approaches have shown great potential for alleviating. Nevertheless, these approaches have certain limitations. They usually do not fully consider the problem of incompatible feature embeddings. Instead, they tend to focus only on the features of new or previous classes and fail to comprehensively consider the entire model. Therefore, we propose a two-stage learning paradigm to solve feature embedding incompatibility problems. Specifically, we retain the previous model and freeze all its parameters in the first stage while dynamically expanding a new module to alleviate feature embedding incompatibility questions. In the second stage, a fusion knowledge distillation approach is used to compress the redundant feature dimensions. Moreover, we propose weight pruning and consolidation approaches to improve the efficiency of the model. Our experimental results obtained on the CIFAR-100, ImageNet-100 and ImageNet-1000 benchmark datasets show that the proposed approaches achieve the best performance among all the compared approaches. For example, on the ImageNet-100 dataset, the maximal accuracy improvement is 5.08%. Code is available at https://github.com/ybyangjing/CIL-FCE.
引用
收藏
页数:13
相关论文
共 55 条
[1]  
Rusu AA, 2016, Arxiv, DOI arXiv:1606.04671
[2]   Memory Aware Synapses: Learning What (not) to Forget [J].
Aljundi, Rahaf ;
Babiloni, Francesca ;
Elhoseiny, Mohamed ;
Rohrbach, Marcus ;
Tuytelaars, Tinne .
COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 :144-161
[3]   Expert Gate: Lifelong Learning with a Network of Experts [J].
Aljundi, Rahaf ;
Chakravarty, Punarjay ;
Tuytelaars, Tinne .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :7120-7129
[4]  
Balan Anoop Korattikara, 2015, Advances in neural information processing systems, P3438
[5]   Class-Incremental Continual Learning Into the eXtended DER-Verse [J].
Boschini, Matteo ;
Bonicelli, Lorenzo ;
Buzzega, Pietro ;
Porrello, Angelo ;
Calderara, Simone .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (05) :5497-5512
[6]   Class-Balanced Loss Based on Effective Number of Samples [J].
Cui, Yin ;
Jia, Menglin ;
Lin, Tsung-Yi ;
Song, Yang ;
Belongie, Serge .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :9260-9269
[7]   A Continual Learning Survey: Defying Forgetting in Classification Tasks [J].
De Lange, Matthias ;
Aljundi, Rahaf ;
Masana, Marc ;
Parisot, Sarah ;
Jia, Xu ;
Leonardis, Ales ;
Slabaugh, Greg ;
Tuytelaars, Tinne .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (07) :3366-3385
[8]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[9]   Learning without Memorizing [J].
Dhar, Prithviraj ;
Singh, Rajat Vikram ;
Peng, Kuan-Chuan ;
Wu, Ziyan ;
Chellappa, Rama .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :5133-5141
[10]   DyTox: Transformers for Continual Learning with DYnamic TOken eXpansion [J].
Douillard, Arthur ;
Rame, Alexandre ;
Couairon, Guillaume ;
Cord, Matthieu .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, :9275-9285