A robust and anti-forgettiable model for class-incremental learning

被引:1
作者
Chen, Jianting [1 ]
Xiang, Yang [1 ]
机构
[1] Tongji Univ, Coll Elect & Informat Engn, 4800 Caoan Highway, Shanghai 201804, Peoples R China
基金
中国国家自然科学基金;
关键词
Class incremental learning; Catastrophic forgetting; Batch normalization; Robust feature representation;
D O I
10.1007/s10489-022-04239-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In many real-world scenarios, neural network models are not always fixed; they are expected to adapt to a dynamic environment and incrementally learn new knowledge. However, catastrophic forgetting is a challenge for incremental learning in neural networks since updating the model parameters to incorporate new knowledge often results in performance degradation on previous tasks. In this paper, we focus on class-incremental learning (CIL) and attempt to mitigate catastrophic forgetting by improving the robustness of neural networks. Specifically, we modify two aspects of the models. First, we argue that plain batch normalization (BN) has a negative effect on CIL. Hence, we propose a variant BN, called noisy batch normalization (NBN), which introduces Gaussian noise to resist the impact of the change in feature distributions and improves feature representation robustness. Second, to address the task-level overfitting problem in CIL, we introduce a decoder-based regularization (DBR) term, which employs a decoder following the feature encoder to reconstruct the input. DBR can avoid overfitting of the current task and provide a distillation loss to retain the knowledge of previous tasks. We design two CIL scenarios and validate our approaches on the CIFAR-100, MiniImageNet, Fashion MNIST, and Omniglot datasets. The results show that the performance of CIL algorithms based on our approach is better than that of the original algorithms, indicating that our approach can enhance the model robustness and help the networks extract anti-forgettable feature representations.
引用
收藏
页码:14128 / 14145
页数:18
相关论文
共 50 条
  • [21] Adaptive adapter routing for long-tailed class-incremental learning
    Qi, Zhi-Hong
    Zhou, Da-Wei
    Yao, Yiran
    Ye, Han-Jia
    Zhan, De-Chuan
    MACHINE LEARNING, 2025, 114 (03)
  • [22] PILoRA: Prototype Guided Incremental LoRA for Federated Class-Incremental Learning
    Guo, Haiyang
    Zhu, Fei
    Liu, Wenzhuo
    Zhang, Xu-Yao
    Liu, Cheng-Lin
    COMPUTER VISION - ECCV 2024, PT LXV, 2025, 15123 : 141 - 159
  • [23] Continual prune-and-select: class-incremental learning with specialized subnetworks
    Dekhovich, Aleksandr
    Tax, David M. J.
    Sluiter, Marcel H. F.
    Bessa, Miguel A.
    APPLIED INTELLIGENCE, 2023, 53 (14) : 17849 - 17864
  • [24] CBCL-PR: A Cognitively Inspired Model for Class-Incremental Learning in Robotics
    Ayub, Ali
    Wagner, Alan R.
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (04) : 2004 - 2013
  • [25] Squeezing More Past Knowledge for Online Class-Incremental Continual Learning
    Da Yu
    Mingyi Zhang
    Mantian Li
    Fusheng Zha
    Junge Zhang
    Lining Sun
    Kaiqi Huang
    IEEE/CAAJournalofAutomaticaSinica, 2023, 10 (03) : 722 - 736
  • [26] Class-Incremental Learning with Topological Schemas of Memory Spaces
    Chang, Xinyuan
    Tao, Xiaoyu
    Hong, Xiaopeng
    Wei, Xing
    Ke, Wei
    Gong, Yihong
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 9719 - 9726
  • [27] Dynamic Task Subspace Ensemble for Class-Incremental Learning
    Zhang, Weile
    He, Yuanjian
    Cong, Yulai
    ARTIFICIAL INTELLIGENCE, CICAI 2023, PT II, 2024, 14474 : 322 - 334
  • [28] Class-Incremental Generalized Zero-Shot Learning
    Zhenfeng Sun
    Rui Feng
    Yanwei Fu
    Multimedia Tools and Applications, 2023, 82 : 38233 - 38247
  • [30] Deep Class-Incremental Learning From Decentralized Data
    Zhang, Xiaohan
    Dong, Songlin
    Chen, Jinjie
    Tian, Qi
    Gong, Yihong
    Hong, Xiaopeng
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (05) : 7190 - 7203