MoBoo: Memory-Boosted Vision Transformer for Class-Incremental Learning

被引:0
|
作者
Ni, Bolin [1 ,2 ]
Nie, Xing [1 ,2 ]
Zhang, Chenghao [1 ,2 ]
Xu, Shixiong [1 ,2 ]
Zhang, Xin [3 ]
Meng, Gaofeng [1 ,2 ,4 ]
Xiang, Shiming [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[3] Beijing Inst Technol, Sch Informat & Elect, Radar Res Lab, Beijing 100081, Peoples R China
[4] HK Inst Sci & Innovat, CAS Ctr Artificial Intelligence & Robot, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Continual learning; class-incremental learning; vision transformer; image recognition;
D O I
10.1109/TCSVT.2024.3417431
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Continual learning strives to acquire knowledge across sequential tasks without forgetting previously assimilated knowledge. Current state-of-the-art methodologies utilize dynamic architectural strategies to increase the network capacity for new tasks. However, these approaches often suffer from a rapid growth in the number of parameters. While some methods introduce an additional network compression stage to address this, they tend to construct complex and hyperparameter-sensitive systems. In this work, we introduce a novel solution to this challenge by proposing Memory-Boosted transformer (MoBoo), instead of conventional architecture expansion and compression. Specifically, we design a memory-augmented attention mechanism by establishing a memory bank where the "key" and "value" linear projections are stored. This memory integration prompts the model to leverage previously learned knowledge, thereby enhancing stability during training at a marginal cost. The memory bank is lightweight and can be easily managed with a straightforward queue. Moreover, to increase the model's plasticity, we design a memory-attentive aggregator, which leverages the cross-attention mechanism to adaptively summarize the image representation from the encoder output that has historical knowledge involved. Extensive experiments on challenging benchmarks demonstrate the effectiveness of our method. For example, on ImageNet-100 under 10 tasks, our method outperforms the current state-of-the-art methods by +3.74% in average accuracy and using fewer parameters.
引用
收藏
页码:11169 / 11183
页数:15
相关论文
共 50 条
  • [21] Dynamic Task Subspace Ensemble for Class-Incremental Learning
    Zhang, Weile
    He, Yuanjian
    Cong, Yulai
    ARTIFICIAL INTELLIGENCE, CICAI 2023, PT II, 2024, 14474 : 322 - 334
  • [22] Memory-Enhanced Confidence Calibration for Class-Incremental Unsupervised Domain Adaptation
    Yu, Jiaping
    Yang, Muli
    Wu, Aming
    Deng, Cheng
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 610 - 621
  • [23] Class-Incremental Generalized Zero-Shot Learning
    Zhenfeng Sun
    Rui Feng
    Yanwei Fu
    Multimedia Tools and Applications, 2023, 82 : 38233 - 38247
  • [25] FOSTER: Feature Boosting and Compression for Class-Incremental Learning
    Wang, Fu-Yun
    Zhou, Da-Wei
    Ye, Han-Jia
    Zhan, De-Chuan
    COMPUTER VISION, ECCV 2022, PT XXV, 2022, 13685 : 398 - 414
  • [26] Mixup-Inspired Video Class-Incremental Learning
    Long, Jinqiang
    Gao, Yizhao
    Lu, Zhiwu
    23RD IEEE INTERNATIONAL CONFERENCE ON DATA MINING, ICDM 2023, 2023, : 1181 - 1186
  • [27] Class-incremental learning with Balanced Embedding Discrimination Maximization
    Wei, Qinglai
    Zhang, Weiqin
    NEURAL NETWORKS, 2024, 179
  • [28] Class-Incremental Generalized Zero-Shot Learning
    Sun, Zhenfeng
    Feng, Rui
    Fu, Yanwei
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (24) : 38233 - 38247
  • [29] Class-Incremental Learning for Baseband Modulation Classification: A Comparison
    Montes, Charles
    Morehouse, Todd
    Zhou, Ruolin
    20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024, 2024, : 1370 - 1375
  • [30] Learning to complement: Relation complementation network for few-shot class-incremental learning
    Wang, Ye
    Wang, Yaxiong
    Zhao, Guoshuai
    Qian, Xueming
    KNOWLEDGE-BASED SYSTEMS, 2023, 282