MoBoo: Memory-Boosted Vision Transformer for Class-Incremental Learning

被引:0
|
作者
Ni, Bolin [1 ,2 ]
Nie, Xing [1 ,2 ]
Zhang, Chenghao [1 ,2 ]
Xu, Shixiong [1 ,2 ]
Zhang, Xin [3 ]
Meng, Gaofeng [1 ,2 ,4 ]
Xiang, Shiming [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[3] Beijing Inst Technol, Sch Informat & Elect, Radar Res Lab, Beijing 100081, Peoples R China
[4] HK Inst Sci & Innovat, CAS Ctr Artificial Intelligence & Robot, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Continual learning; class-incremental learning; vision transformer; image recognition;
D O I
10.1109/TCSVT.2024.3417431
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Continual learning strives to acquire knowledge across sequential tasks without forgetting previously assimilated knowledge. Current state-of-the-art methodologies utilize dynamic architectural strategies to increase the network capacity for new tasks. However, these approaches often suffer from a rapid growth in the number of parameters. While some methods introduce an additional network compression stage to address this, they tend to construct complex and hyperparameter-sensitive systems. In this work, we introduce a novel solution to this challenge by proposing Memory-Boosted transformer (MoBoo), instead of conventional architecture expansion and compression. Specifically, we design a memory-augmented attention mechanism by establishing a memory bank where the "key" and "value" linear projections are stored. This memory integration prompts the model to leverage previously learned knowledge, thereby enhancing stability during training at a marginal cost. The memory bank is lightweight and can be easily managed with a straightforward queue. Moreover, to increase the model's plasticity, we design a memory-attentive aggregator, which leverages the cross-attention mechanism to adaptively summarize the image representation from the encoder output that has historical knowledge involved. Extensive experiments on challenging benchmarks demonstrate the effectiveness of our method. For example, on ImageNet-100 under 10 tasks, our method outperforms the current state-of-the-art methods by +3.74% in average accuracy and using fewer parameters.
引用
收藏
页码:11169 / 11183
页数:15
相关论文
共 50 条
  • [41] A robust and anti-forgettiable model for class-incremental learning
    Jianting Chen
    Yang Xiang
    Applied Intelligence, 2023, 53 : 14128 - 14145
  • [42] Class-Incremental Continual Learning Into the eXtended DER-Verse
    Boschini, Matteo
    Bonicelli, Lorenzo
    Buzzega, Pietro
    Porrello, Angelo
    Calderara, Simone
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (05) : 5497 - 5512
  • [43] Towards Effective Authorship Attribution: Integrating Class-Incremental Learning
    Rahgouy, Mostafa
    Giglou, Hamed Babaei
    Tabassum, Mehnaz
    Feng, Dongji
    Das, Amit
    Rahgooy, Taher
    Dozier, Gerry
    Seals, Cheryl D.
    2024 IEEE 6TH INTERNATIONAL CONFERENCE ON COGNITIVE MACHINE INTELLIGENCE, COGMI, 2024, : 56 - 65
  • [44] Online Class-Incremental Learning in Image Classification Based on Attention
    Du, Baoyu
    Wei, Zhonghe
    Cheng, Jinyong
    Lv, Guohua
    Dai, Xiaoyu
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VII, 2024, 14431 : 487 - 499
  • [45] A robust and anti-forgettiable model for class-incremental learning
    Chen, Jianting
    Xiang, Yang
    APPLIED INTELLIGENCE, 2023, 53 (11) : 14128 - 14145
  • [46] Anchor Assisted Experience Replay for Online Class-Incremental Learning
    Lin, Huiwei
    Feng, Shanshan
    Li, Xutao
    Li, Wentao
    Ye, Yunming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (05) : 2217 - 2232
  • [47] Preserving Locality in Vision Transformers for Class Incremental Learning
    Zheng, Bowen
    Zhou, Wei
    Ye, Han-Jia
    Zhan, De-Chuan
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1157 - 1162
  • [48] Adaptive adapter routing for long-tailed class-incremental learning
    Qi, Zhi-Hong
    Zhou, Da-Wei
    Yao, Yiran
    Ye, Han-Jia
    Zhan, De-Chuan
    MACHINE LEARNING, 2025, 114 (03)
  • [49] Enhancing Class-Incremental Learning for Image Classification via Bidirectional Transport and Selective Momentum
    Fu, Feifei
    Gao, Yizhao
    Lu, Zhiwu
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 175 - 183
  • [50] CLASS-INCREMENTAL LEARNING FOR REMOTE SENSING IMAGES BASED ON KNOWLEDGE DISTILLATION
    Song, Jingduo
    Jia, Hecheng
    Xu, Feng
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 5026 - 5028