Efficient Statistical Sampling Adaptation for Exemplar-Free Class Incremental Learning

被引:1
|
作者
Cheng, De [1 ]
Zhao, Yuxin [1 ]
Wang, Nannan [1 ]
Li, Guozhang [2 ]
Zhang, Dingwen
Gao, Xinbo
机构
[1] Xidian Univ, Sch Telecommun Engn, State Key Lab Integrated Serv Networks, Xian 710071, Shaanxi, Peoples R China
[2] Xidian Univ, Sch Elect Engn, State Key Lab Integrated Serv Networks, Xian 710071, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Exemplar-free; class incremental learning; catastrophic forgetting; feature statistics; adaptation;
D O I
10.1109/TCSVT.2024.3421587
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep learning systems typically suffer from catastrophic forgetting of old knowledge when learning from new data continually. Recently, various class incremental learning (CIL) methods have been proposed to address this issue, and some approaches achieve promising performances by relying on rehearsing the training data of previous tasks. However, storing data from previous tasks would encounter data privacy and memory issues in real-world applications. In this paper, we propose a statistical sampling adaptation method for efficient Exemplar-Free Class-Incremental Learning (EFCIL). Here, instead of preserving the images/features themselves of previous tasks/classes, we store image feature statistics from previous classes to maintain the decision boundary, which is memory-efficient and much semantic-representative. When utilizing the old-class feature statistics, we build a statistical feature adaptation network (SFAN) with a manifold consistency regularization and then train it in a transductive learning paradigm, which can map the outdated statistics onto the current feature space to facilitate a compatible and balanced classifier training subsequently. In this way, the final classifier can be jointly optimized with all the old-class features projected by SFAN and current new-class features, thus alleviating the classification bias problem in EFCIL. Experimental results greatly demonstrate the effectiveness of the proposed method, achieving superior performances than state-of-the-art approaches. Our source code is released in https://github.com/yxzhcv/ESSA-EFCIL.
引用
收藏
页码:11451 / 11463
页数:13
相关论文
共 21 条
  • [1] Rotation Augmented Distillation for Exemplar-Free Class Incremental Learning with Detailed Analysis
    Chen, Xiuwei
    Chang, Xiaobin
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT IV, 2024, 14428 : 27 - 38
  • [2] Adaptive Margin Global Classifier for Exemplar-Free Class-Incremental Learning
    Yao, Zhongren
    Chang, Xiaobin
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT 1, 2025, 15031 : 476 - 489
  • [3] Representation Robustness and Feature Expansion for Exemplar-Free Class-Incremental Learning
    Luo, Yong
    Ge, Hongwei
    Liu, Yuxuan
    Wu, Chunguo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (07) : 5306 - 5320
  • [4] Self-distilled Knowledge Delegator for Exemplar-free Class Incremental Learning
    Ye, Fanfan
    Ma, Liang
    Zhong, Qiaoyong
    Xie, Di
    Pu, Shiliang
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [5] Adaptive Knowledge Matching for Exemplar-Free Class-Incremental Learning
    Chen, Runhang
    Jing, Xiao-Yuan
    Chen, Haowen
    PATTERN RECOGNITION AND COMPUTER VISION, PT III, PRCV 2024, 2025, 15033 : 289 - 303
  • [6] FRMM: Feature Reprojection for Exemplar-Free Class-Incremental Learning
    Wang, Hao
    Chen, Jing
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT III, ICIC 2024, 2024, 14864 : 251 - 263
  • [7] Less confidence, less forgetting: Learning with a humbler teacher in exemplar-free Class-Incremental learning
    Gao, Zijian
    Xu, Kele
    Zhuang, Huiping
    Liu, Li
    Mao, Xinjun
    Ding, Bo
    Feng, Dawei
    Wang, Huaimin
    NEURAL NETWORKS, 2024, 179
  • [8] Multi-view prototype balance and temporary proxy constraint for exemplar-free class-incremental learning
    Tian, Heng
    Zhang, Qian
    Wang, Zhe
    Zhang, Yu
    Xu, Xinlei
    Fu, Zhiling
    APPLIED INTELLIGENCE, 2025, 55 (05)
  • [9] EXEMPLAR-FREE ONLINE CONTINUAL LEARNING
    He, Jiangpeng
    Zhu, Fengqing
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 541 - 545
  • [10] Exemplar-Free Continual Learning of Vision Transformers via Gated Class-Attention and Cascaded Feature Drift Compensation
    Cotogni, Marco
    Yang, Fei
    Cusano, Claudio
    Bagdanov, Andrew D.
    van de Weijer, Joost
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025, : 4571 - 4589