Preserving Locality in Vision Transformers for Class Incremental Learning

被引:3
|
作者
Zheng, Bowen [1 ]
Zhou, Wei [1 ]
Ye, Han-Jia [1 ]
Zhan, De-Chuan [1 ]
机构
[1] Nanjing Univ, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
来源
2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME | 2023年
基金
国家重点研发计划;
关键词
Class Incremental Learning; Vision Transformer;
D O I
10.1109/ICME55011.2023.00202
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning new classes without forgetting is crucial for real-world applications for a classification model. Vision Transformers (ViT) recently achieve remarkable performance in Class Incremental Learning (CIL). Previous works mainly focus on block design and model expansion for ViTs. However, in this paper, we find that when the ViT is incrementally trained, the attention layers gradually lose concentration on local features. We call this interesting phenomenon as Locality Degradation in ViTs for CIL. Since the low-level local information is crucial to the transferability of the representation, it is beneficial to preserve the locality in attention layers. In this paper, we encourage the model to preserve more local information as the training procedure goes on and devise a Locality-Preserved Attention (LPA) layer to emphasize the importance of local features. Specifically, we incorporate the local information directly into the vanilla attention and control the initial gradients of the vanilla attention by weighting it with a small initial value. Extensive experiments show that the representations facilitated by LPA capture more low-level general information which is easier to transfer to follow-up tasks. The improved model gets consistently better performance on CIFAR100 and ImageNet100. The source code is available at https://github.com/bwnzheng/LPA_ICME2023.
引用
收藏
页码:1157 / 1162
页数:6
相关论文
共 50 条
  • [1] MoBoo: Memory-Boosted Vision Transformer for Class-Incremental Learning
    Ni, Bolin
    Nie, Xing
    Zhang, Chenghao
    Xu, Shixiong
    Zhang, Xin
    Meng, Gaofeng
    Xiang, Shiming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (11) : 11169 - 11183
  • [2] Reformulating Classification as Image-Class Matching for Class Incremental Learning
    Hu, Yusong
    Liang, Zichen
    Liu, Xialei
    Hou, Qibin
    Cheng, Ming-Ming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (01) : 811 - 822
  • [3] Class-Incremental Learning: A Survey
    Zhou, Da-Wei
    Wang, Qi-Wei
    Qi, Zhi-Hong
    Ye, Han-Jia
    Zhan, De-Chuan
    Liu, Ziwei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 9851 - 9873
  • [4] Rethinking class orders and transferability in class incremental learning
    He, Chen
    Wang, Ruiping
    Chen, Xilin
    PATTERN RECOGNITION LETTERS, 2022, 161 : 67 - 73
  • [5] Deep Learning for Class-Incremental Learning: A Survey
    Zhou D.-W.
    Wang F.-Y.
    Ye H.-J.
    Zhan D.-C.
    Jisuanji Xuebao/Chinese Journal of Computers, 2023, 46 (08): : 1577 - 1605
  • [6] ISM-Net: Mining incremental semantics for class incremental learning
    Qiu, Zihuan
    Xu, Linfeng
    Wang, Zhichuan
    Wu, Qingbo
    Meng, Fanman
    Li, Hongliang
    NEUROCOMPUTING, 2023, 523 : 130 - 143
  • [7] Multi-view class incremental learning
    Li, Depeng
    Wang, Tianqi
    Chen, Junwei
    Kawaguchi, Kenji
    Lian, Cheng
    Zeng, Zhigang
    INFORMATION FUSION, 2024, 102
  • [8] CLASS INCREMENTAL LEARNING FOR VIDEO ACTION CLASSIFICATION
    Ma, Jiawei
    Tao, Xiaoyu
    Ma, Jianxing
    Hong, Xiaopeng
    Gong, Yihong
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 504 - 508
  • [9] PARALLEL GRADIENT BLEND FOR CLASS INCREMENTAL LEARNING
    Zhao, Yunlong
    Deng, Xiaoheng
    Pei, Xinjun
    Chen, Xuechen
    Li, Deng
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1220 - 1224
  • [10] Knowledge aggregation networks for class incremental learning
    Fu, Zhiling
    Wang, Zhe
    Xu, Xinlei
    Li, Dongdong
    Yang, Hai
    PATTERN RECOGNITION, 2023, 137