Knowledge aggregation networks for class incremental learning

被引:22
作者
Fu, Zhiling [1 ,2 ]
Wang, Zhe [1 ,2 ]
Xu, Xinlei [1 ,2 ]
Li, Dongdong [2 ]
Yang, Hai [2 ]
机构
[1] East China Univ Sci & Technol, Key Lab Smart Mfg Energy Chem Proc, Minist Educ, Shanghai 200237, Peoples R China
[2] East China Univ Sci & Technol, Dept Comp Sci & Engn, Shanghai 200237, Peoples R China
关键词
Class incremental learning; Catastrophic forgetting; Dual-branch network; Knowledge aggregation; Model compression; CONSOLIDATION;
D O I
10.1016/j.patcog.2023.109310
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most existing class incremental learning methods rely on storing old exemplars to avoid catastrophic forgetting. However, these methods inevitably face the gradient conflict problem, the inherent conflict between new streaming knowledge and existing knowledge in the gradient direction. To alleviate gradient conflict, this paper reuses the previous knowledge and expands the branch to accommodate new concepts instead of fine-tuning the original models. Specifically, this paper designs a novel dual-branch network called Knowledge Aggregation Networks. The previously trained model is frozen as a branch to retain existing knowledge, and a consistent trainable network is constructed as the other branch to learn new concepts. An adaptive feature fusion module is adopted to dynamically balance the two branches' information during training. Moreover, a model compression stage maintains the dual-branch structure. Extensive experiments on CIFAR-10 0, ImageNet-Sub, and ImageNet show that our method significantly outperforms the other methods and effectively balances stability and plasticity. & COPY; 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页数:12
相关论文
共 50 条
[21]   KABI: Class-Incremental Learning via knowledge Amalgamation and Batch Identification [J].
Li, Caixia ;
Xu, Wenhua ;
Si, Xizhu ;
Song, Ping .
2021 5TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE (ICIAI 2021), 2021, :170-176
[22]   PKI: Prior knowledge-infused neural network for few-shot class-incremental learning [J].
Bao, Kexin ;
Lin, Fanzhao ;
Wang, Zichen ;
Li, Yong ;
Zeng, Dan ;
Ge, Shiming .
NEURAL NETWORKS, 2025, 192
[23]   Semantic Bridging and Feature Anchoring for Class Incremental Learning [J].
Wu, Kanghui ;
Guo, Dongyan .
2024 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME 2024, 2024,
[24]   Feature expansion and enhanced compression for class incremental learning [J].
Ferdinand, Quentin ;
Clement, Benoit ;
Papadakis, Panagiotis ;
Oliveau, Quentin ;
Le Chenadec, Gilles .
NEUROCOMPUTING, 2025, 634
[25]   A Broad Neural Network Structure for Class Incremental Learning [J].
Liu, Wenzhang ;
Yang, Haiqin ;
Sun, Yuewen ;
Sun, Changyin .
ADVANCES IN NEURAL NETWORKS - ISNN 2018, 2018, 10878 :229-238
[26]   DYNAMIC REPLAY TRAINING FOR CLASS-INCREMENTAL LEARNING [J].
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, :5915-5919
[27]   APM: Adaptive parameter multiplexing for class incremental learning [J].
Gao, Jinghan ;
Xie, Tao ;
Li, Ruifeng ;
Wang, Ke ;
Zhao, Lijun .
EXPERT SYSTEMS WITH APPLICATIONS, 2024, 258
[28]   KiCi: A Knowledge Importance Based Class Incremental Learning Method forWearable Activity Recognition [J].
Guo, Shuai ;
Gu, Yang ;
Wen, Shijie ;
Ma, Yuan ;
Chen, Yiqiang ;
Wang, Jiwei ;
Hu, Chunyu .
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, :646-655
[29]   Knowledge Representation by Generic Models for Few-Shot Class-Incremental Learning [J].
Chen, Xiaodong ;
Jiang, Weijie ;
Huang, Zhiyong ;
Su, Jiangwen ;
Yu, Yuanlong .
ADVANCES IN NATURAL COMPUTATION, FUZZY SYSTEMS AND KNOWLEDGE DISCOVERY, ICNC-FSKD 2022, 2023, 153 :1237-1247
[30]   Self-distilled Knowledge Delegator for Exemplar-free Class Incremental Learning [J].
Ye, Fanfan ;
Ma, Liang ;
Zhong, Qiaoyong ;
Xie, Di ;
Pu, Shiliang .
2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,