APM: Adaptive parameter multiplexing for class incremental learning

被引:0
作者
Gao, Jinghan [1 ]
Xie, Tao [1 ]
Li, Ruifeng [1 ]
Wang, Ke [1 ]
Zhao, Lijun [1 ]
机构
[1] Harbin Inst Technol, State Key Lab Robot & Syst, Harbin 150006, Peoples R China
关键词
Incremental learning; Deep learning; Image classification; Class incremental learning; Catastrophic forgetting;
D O I
10.1016/j.eswa.2024.125135
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent developments within the domain of image classification, deep neural networks (DNNs) have attracted considerable scholarly interest and have been extensively trained using data in closed environments. Such training methodologies contrast sharply with the inherently open, progressive, and adaptive processes of the natural visual system, leading to emergent challenges. Among these, catastrophic forgetting is notable, where the network acquisition of new class information precipitates the erosion of previously established knowledge. Additionally, the network encounters the stability-plasticity dilemma, necessitating a delicate equilibrium between assimilating novel classes and retaining existing ones. To address these issues, we propose a novel incremental learning model, termed Adaptive Parameter Multiplexing (APM), which incorporates a cross-class parameter adaptive incremental strategy. Central to our methodology is the conceptualization of parameter multiplexing or incremental as a learnable optimization problem, enabling the model to autonomously evaluate and decide on the necessity for parameter adjustment throughout its training lifecycle. This framework is designed to enhance the ability of the network to extract features for new class categories effectively through incremental parameters while simultaneously employing parameter multiplexing to augment storage optimization. Our model is underpinned by a dual strategy of coarse-grained and fine-grained parameter multiplexing, guided by a learnable score that dynamically assesses the appropriateness of parameter multiplexing versus incremental updates, facilitating an optimized balance for incremental model performance and storage. In addition, we have integrated a novel regularization loss mechanism for the learnable score to optimize storage efficiency. The effectiveness of APM is empirically validated through rigorous testing on benchmark datasets, including ImageNet100, CIFAR100, CIFAR10, and CUB200. The experimental outcomes indicate that, with a trace amount of parameter increase, our model achieves significant enhancements in classification performance across both new and previously established classes, thereby surpassing existing benchmarks set by state-of-the-art algorithms in the field.
引用
收藏
页数:10
相关论文
共 54 条
  • [1] Aljundi R, 2019, ADV NEUR IN, V32
  • [2] A crowdsourcing-based incremental learning framework for automated essays scoring
    Bai, Huanyu
    Hui, Siu Cheung
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [3] IL2M: Class Incremental Learning With Dual Memory
    Belouadah, Eden
    Popescu, Adrian
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 583 - 592
  • [4] Chaudhry A, 2019, Arxiv, DOI arXiv:1812.00420
  • [5] Chaudhry A, 2021, AAAI CONF ARTIF INTE, V35, P6993
  • [6] Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence
    Chaudhry, Arslan
    Dokania, Puneet K.
    Ajanthan, Thalaiyasingam
    Torr, Philip H. S.
    [J]. COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 : 556 - 572
  • [7] Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
  • [8] Learning without Memorizing
    Dhar, Prithviraj
    Singh, Rajat Vikram
    Peng, Kuan-Chuan
    Wu, Ziyan
    Chellappa, Rama
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 5133 - 5141
  • [9] Douillard Arthur, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12365), P86, DOI 10.1007/978-3-030-58565-5_6
  • [10] DyTox: Transformers for Continual Learning with DYnamic TOken eXpansion
    Douillard, Arthur
    Rame, Alexandre
    Couairon, Guillaume
    Cord, Matthieu
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9275 - 9285