Less confidence, less forgetting: Learning with a humbler teacher in exemplar-free Class-Incremental learning

被引:1
|
作者
Gao, Zijian [1 ,2 ]
Xu, Kele [1 ,2 ]
Zhuang, Huiping [3 ]
Liu, Li [1 ,2 ,4 ]
Mao, Xinjun [1 ,2 ]
Ding, Bo [1 ,2 ]
Feng, Dawei [1 ,2 ]
Wang, Huaimin [1 ,2 ]
机构
[1] Natl Univ Def Technol, Changsha 410000, Peoples R China
[2] State Key Lab Complex & Crit Software Environm, Changsha 410000, Peoples R China
[3] South China Univ Technol, Guangzhou 510000, Peoples R China
[4] Univ Oulu, Oulu, Finland
基金
中国国家自然科学基金;
关键词
Exemplar-free Class-Incremental learning; Catastrophic forgetting; Knowledge distillation; Checkpoint model;
D O I
10.1016/j.neunet.2024.106513
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Class-Incremental learning (CIL) is challenging due to catastrophic forgetting (CF), which escalates in exemplarfree scenarios. To mitigate CF, Knowledge Distillation (KD), which leverages old models as teacher models, has been widely employed in CIL. However, based on a case study, our investigation reveals that the teacher model exhibits over-confidence in unseen new samples. In this article, we conduct empirical experiments and provide theoretical analysis to investigate the over-confident phenomenon and the impact of KD in exemplar-free CIL, where access to old samples is unavailable. Building on our analysis, we propose a novel approach, Learning with Humbler Teacher, by systematically selecting an appropriate checkpoint model as a humbler teacher to mitigate CF. Furthermore, we explore utilizing the nuclear norm to obtain an appropriate temporal ensemble to enhance model stability. Notably, LwHT outperforms the state-of-the-art approach by a significant margin of 10.41%, 6.56%, and 4.31% in various settings while demonstrating superior model plasticity.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Memory Efficient Class-Incremental Learning for Image Classification
    Zhao, Hanbin
    Wang, Hui
    Fu, Yongjian
    Wu, Fei
    Li, Xi
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (10) : 5966 - 5977
  • [22] KABI: Class-Incremental Learning via knowledge Amalgamation and Batch Identification
    Li, Caixia
    Xu, Wenhua
    Si, Xizhu
    Song, Ping
    2021 5TH INTERNATIONAL CONFERENCE ON INNOVATION IN ARTIFICIAL INTELLIGENCE (ICIAI 2021), 2021, : 170 - 176
  • [23] Prompt-based learning for few-shot class-incremental learning
    Yuan, Jicheng
    Chen, Hang
    Tian, Songsong
    Li, Wenfa
    Li, Lusi
    Ning, Enhao
    Zhang, Yugui
    ALEXANDRIA ENGINEERING JOURNAL, 2025, 120 : 287 - 295
  • [24] Rethinking Few-Shot Class-Incremental Learning: Learning from Yourself
    Tang, Yu-Ming
    Peng, Yi-Xing
    Meng, Jingke
    Zheng, Wei-Shi
    COMPUTER VISION - ECCV 2024, PT LXI, 2025, 15119 : 108 - 128
  • [25] A robust and anti-forgettiable model for class-incremental learning
    Jianting Chen
    Yang Xiang
    Applied Intelligence, 2023, 53 : 14128 - 14145
  • [26] A robust and anti-forgettiable model for class-incremental learning
    Chen, Jianting
    Xiang, Yang
    APPLIED INTELLIGENCE, 2023, 53 (11) : 14128 - 14145
  • [27] General Federated Class-Incremental Learning With Lightweight Generative Replay
    Chen, Yuanlu
    Tan, Alysa Ziying
    Feng, Siwei
    Yu, Han
    Deng, Tao
    Zhao, Libang
    Wu, Feng
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (20): : 33927 - 33939
  • [28] Class-Incremental Learning: Survey and Performance Evaluation on Image Classification
    Masana, Marc
    Liu, Xialei
    Twardowski, Bartlomiej
    Menta, Mikel
    Bagdanov, Andrew D.
    van de Weijer, Joost
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (05) : 5513 - 5533
  • [29] Continual prune-and-select: class-incremental learning with specialized subnetworks
    Aleksandr Dekhovich
    David M.J. Tax
    Marcel H.F Sluiter
    Miguel A. Bessa
    Applied Intelligence, 2023, 53 : 17849 - 17864
  • [30] Adaptive adapter routing for long-tailed class-incremental learning
    Qi, Zhi-Hong
    Zhou, Da-Wei
    Yao, Yiran
    Ye, Han-Jia
    Zhan, De-Chuan
    MACHINE LEARNING, 2025, 114 (03)