Knowledge aggregation networks for class incremental learning

被引:22
作者
Fu, Zhiling [1 ,2 ]
Wang, Zhe [1 ,2 ]
Xu, Xinlei [1 ,2 ]
Li, Dongdong [2 ]
Yang, Hai [2 ]
机构
[1] East China Univ Sci & Technol, Key Lab Smart Mfg Energy Chem Proc, Minist Educ, Shanghai 200237, Peoples R China
[2] East China Univ Sci & Technol, Dept Comp Sci & Engn, Shanghai 200237, Peoples R China
关键词
Class incremental learning; Catastrophic forgetting; Dual-branch network; Knowledge aggregation; Model compression; CONSOLIDATION;
D O I
10.1016/j.patcog.2023.109310
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most existing class incremental learning methods rely on storing old exemplars to avoid catastrophic forgetting. However, these methods inevitably face the gradient conflict problem, the inherent conflict between new streaming knowledge and existing knowledge in the gradient direction. To alleviate gradient conflict, this paper reuses the previous knowledge and expands the branch to accommodate new concepts instead of fine-tuning the original models. Specifically, this paper designs a novel dual-branch network called Knowledge Aggregation Networks. The previously trained model is frozen as a branch to retain existing knowledge, and a consistent trainable network is constructed as the other branch to learn new concepts. An adaptive feature fusion module is adopted to dynamically balance the two branches' information during training. Moreover, a model compression stage maintains the dual-branch structure. Extensive experiments on CIFAR-10 0, ImageNet-Sub, and ImageNet show that our method significantly outperforms the other methods and effectively balances stability and plasticity. & COPY; 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页数:12
相关论文
共 50 条
[31]   Class incremental learning of remote sensing images based on class similarity distillation [J].
Shen, Mingge ;
Chen, Dehu ;
Hu, Silan ;
Xu, Gang .
PEERJ COMPUTER SCIENCE, 2023, 9
[32]   Rethinking class orders and transferability in class incremental learning [J].
He, Chen ;
Wang, Ruiping ;
Chen, Xilin .
PATTERN RECOGNITION LETTERS, 2022, 161 :67-73
[33]   A class-incremental learning approach for learning feature-compatible embeddings [J].
An, Hongchao ;
Yang, Jing ;
Zhang, Xiuhua ;
Ruan, Xiaoli ;
Wu, Yuankai ;
Li, Shaobo ;
Hu, Jianjun .
NEURAL NETWORKS, 2024, 180
[34]   Class-Incremental Learning of Convolutional Neural Networks Based on Double Consolidation Mechanism [J].
Jin, Leilei ;
Liang, Hong ;
Yang, Changsheng .
IEEE ACCESS, 2020, 8 :172553-172562
[35]   DCIGAN: A Distributed Class-Incremental Learning Method Based on Generative Adversarial Networks [J].
Guan, Hongtao ;
Wang, Yijie ;
Ma, Xingkong ;
Li, Yongmou .
2019 IEEE INTL CONF ON PARALLEL & DISTRIBUTED PROCESSING WITH APPLICATIONS, BIG DATA & CLOUD COMPUTING, SUSTAINABLE COMPUTING & COMMUNICATIONS, SOCIAL COMPUTING & NETWORKING (ISPA/BDCLOUD/SOCIALCOM/SUSTAINCOM 2019), 2019, :768-775
[36]   A Class Incremental Learning and Memory Fusion Method Using Random Weight Neural Networks [J].
Li, De-Peng ;
Zeng, Zhi-Gang .
Zidonghua Xuebao/Acta Automatica Sinica, 2023, 49 (12) :2467-2480
[37]   A survey on few-shot class-incremental learning [J].
Tian, Songsong ;
Li, Lusi ;
Li, Weijun ;
Ran, Hang ;
Ning, Xin ;
Tiwari, Prayag .
NEURAL NETWORKS, 2024, 169 :307-324
[38]   FeTT: Class-Incremental Learning with Feature Transformation Tuning [J].
Qiang, Sunyuan ;
Liang, Yanyan .
MATHEMATICS, 2025, 13 (07)
[39]   PROMPTING TO PROMPT FOR REHEARSAL-FREE CLASS INCREMENTAL LEARNING [J].
Zhao, Guangzhi ;
Hou, Yuting ;
Mu, Kedian .
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, :5405-5409
[40]   Privacy-Preserving Federated Class-Incremental Learning [J].
Xiao, Jue ;
Tang, Xueming ;
Lu, Songfeng .
IEEE TRANSACTIONS ON MACHINE LEARNING IN COMMUNICATIONS AND NETWORKING, 2024, 2 :150-168