MoBoo: Memory-Boosted Vision Transformer for Class-Incremental Learning

被引:0
|
作者
Ni, Bolin [1 ,2 ]
Nie, Xing [1 ,2 ]
Zhang, Chenghao [1 ,2 ]
Xu, Shixiong [1 ,2 ]
Zhang, Xin [3 ]
Meng, Gaofeng [1 ,2 ,4 ]
Xiang, Shiming [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Automat, State Key Lab Multimodal Artificial Intelligence, Beijing 100190, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing 100049, Peoples R China
[3] Beijing Inst Technol, Sch Informat & Elect, Radar Res Lab, Beijing 100081, Peoples R China
[4] HK Inst Sci & Innovat, CAS Ctr Artificial Intelligence & Robot, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Continual learning; class-incremental learning; vision transformer; image recognition;
D O I
10.1109/TCSVT.2024.3417431
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Continual learning strives to acquire knowledge across sequential tasks without forgetting previously assimilated knowledge. Current state-of-the-art methodologies utilize dynamic architectural strategies to increase the network capacity for new tasks. However, these approaches often suffer from a rapid growth in the number of parameters. While some methods introduce an additional network compression stage to address this, they tend to construct complex and hyperparameter-sensitive systems. In this work, we introduce a novel solution to this challenge by proposing Memory-Boosted transformer (MoBoo), instead of conventional architecture expansion and compression. Specifically, we design a memory-augmented attention mechanism by establishing a memory bank where the "key" and "value" linear projections are stored. This memory integration prompts the model to leverage previously learned knowledge, thereby enhancing stability during training at a marginal cost. The memory bank is lightweight and can be easily managed with a straightforward queue. Moreover, to increase the model's plasticity, we design a memory-attentive aggregator, which leverages the cross-attention mechanism to adaptively summarize the image representation from the encoder output that has historical knowledge involved. Extensive experiments on challenging benchmarks demonstrate the effectiveness of our method. For example, on ImageNet-100 under 10 tasks, our method outperforms the current state-of-the-art methods by +3.74% in average accuracy and using fewer parameters.
引用
收藏
页码:11169 / 11183
页数:15
相关论文
共 50 条
  • [31] Few-Shot Class-Incremental Learning for Classification and Object Detection: A Survey
    Zhang, Jinghua
    Liu, Li
    Silven, Olli
    Pietikainen, Matti
    Hu, Dewen
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (04) : 2924 - 2945
  • [32] Federated Class-Incremental Learning With Dynamic Feature Extractor Fusion
    Lu, Yanyan
    Yang, Lei
    Chen, Hao-Rui
    Cao, Jiannong
    Lin, Wanyu
    Long, Saiqin
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (12) : 12969 - 12982
  • [33] A class-incremental learning approach for learning feature-compatible embeddings
    An, Hongchao
    Yang, Jing
    Zhang, Xiuhua
    Ruan, Xiaoli
    Wu, Yuankai
    Li, Shaobo
    Hu, Jianjun
    NEURAL NETWORKS, 2024, 180
  • [34] Visual-Textual Attribute Learning for Class-Incremental Facial Expression Recognition
    Lv, Yuanling
    Huang, Guangyu
    Yan, Yan
    Xue, Jing-Hao
    Chen, Si
    Wang, Hanzi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 8038 - 8051
  • [35] DSS: A Diverse Sample Selection Method to Preserve Knowledge in Class-Incremental Learning
    Nokhwal, Sahil
    Kumar, Nirman
    2023 10TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING & MACHINE INTELLIGENCE, ISCMI, 2023, : 178 - 182
  • [36] Class-Incremental Learning for Recognition of Complex-Valued Signals
    Fan, Zhaoyu
    Tu, Ya
    Lin, Yun
    Shi, Qingjiang
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2024, 10 (02) : 417 - 428
  • [37] An empirical evaluation of tinyML architectures for Class-Incremental Continual Learning
    Tremonti, Matteo
    Pezze, Davide Dalle
    Paissan, Francesco
    Farella, Elisabetta
    Susto, Gian Antonio
    2024 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS AND OTHER AFFILIATED EVENTS, PERCOM WORKSHOPS, 2024, : 690 - 695
  • [38] Exemplar-Supported Representation for Effective Class-Incremental Learning
    Guo, Lei
    Xie, Gang
    Xu, Xinying
    Ren, Jinchang
    IEEE ACCESS, 2020, 8 : 51276 - 51284
  • [39] On the Approximation Risk of Few-Shot Class-Incremental Learning
    Wang, Xuan
    Ji, Zhong
    Liu, Xiyao
    Pang, Yanwei
    Han, Jungong
    COMPUTER VISION - ECCV 2024, PT LI, 2025, 15109 : 162 - 178
  • [40] CLASS-INCREMENTAL LEARNING FOR MULTI-LABEL AUDIO CLASSIFICATION
    Mulimani, Manjunath
    Mesaros, Annamaria
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, ICASSP 2024, 2024, : 916 - 920