Class similarity weighted knowledge distillation for few shot incremental learning

被引:2
作者
Akmel, Feidu [1 ]
Meng, Fanman [1 ]
Wu, Qingbo [1 ]
Chen, Shuai [1 ]
Zhang, Runtong [1 ]
Assefa, Maregu [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu, Peoples R China
关键词
Knowledge distillation; Semantic information; Few shot; Incremental learning;
D O I
10.1016/j.neucom.2024.127587
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few -shot class incremental learning illustrates the challenges of learning new concepts, where the learner can access only a small sample per concept. The standard incremental learning techniques cannot be applied directly because of the small number of samples for training. Moreover, catastrophic forgetting is the propensity of an Artificial Neural Network to fully and abruptly forget previously learned knowledge upon learning new knowledge. This problem happens due to a lack of supervision in older classes or an imbalance between the old and new classes. In this work, we propose a new distillation structure to tackle the forgetting and overfitting issues. Particularly, we suggest a dual distillation module that adaptably draws knowledge from two different but complementary teachers. The first teacher is the base model, which has been trained on large class data, and the second teacher is the updated model from the previous K-1 session, which contains the modified knowledge of previously observed new classes. Thus, the first teacher can reduce overfitting issues by transferring the knowledge obtained from the base classes to the new classes. While the second teacher can reduce knowledge forgetting by distilling knowledge from the previous model. Additionally, we use semantic information as word embedding to facilitate the distillation process. To align visual and semantic vectors, we used the attention mechanism of the embedding of visual data. With extensive experiments on different data sets such as Mini-ImageNet, CIFAR100, and CUB200, our model shows state-of-the-art performance compared to the existing few shot incremental learning methods.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Enhancing few-shot class-incremental learning through prototype optimizationEnhancing few-shot class-incremental learning...M. Jiang et al.
    Mengjuan Jiang
    Jiaqing Fan
    Fanzhang Li
    Applied Intelligence, 2025, 55 (10)
  • [22] Knowledge distillation for incremental learning in semantic segmentation
    Michieli, Umberto
    Zanuttigh, Pietro
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2021, 205
  • [23] Overcomplete-to-sparse representation learning for few-shot class-incremental learning
    Fu, Mengying
    Liu, Binghao
    Ma, Tianren
    Ye, Qixiang
    MULTIMEDIA SYSTEMS, 2024, 30 (02)
  • [24] Overcomplete-to-sparse representation learning for few-shot class-incremental learning
    Fu Mengying
    Liu Binghao
    Ma Tianren
    Ye Qixiang
    Multimedia Systems, 2024, 30
  • [25] FEW-SHOT CLASS INCREMENTAL LEARNING FOR HYPERSPECTRAL IMAGE CLASSIFICATION BASED ON CONSTANTLY UPDATED CLASSIFIER
    Zhao, Yuanyuan
    Ha, Lin
    Wang, Hongyu
    Ma, Xiaorui
    2022 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2022), 2022, : 1376 - 1379
  • [26] Rethinking Self-Supervision for Few-Shot Class-Incremental Learning
    Zhao, Linglan
    Lu, Jing
    Cheng, Zhanzhan
    Liu, Duo
    Fang, Xiangzhong
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 726 - 731
  • [27] Semantics-Driven Generative Replay for Few-Shot Class Incremental Learning
    Agarwal, Aishwarya
    Banerjee, Biplab
    Cuzzolin, Fabio
    Chaudhuri, Subhasis
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5246 - 5254
  • [28] Few-Shot Class-Incremental Learning for Classification and Object Detection: A Survey
    Zhang, Jinghua
    Liu, Li
    Silven, Olli
    Pietikainen, Matti
    Hu, Dewen
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (04) : 2924 - 2945
  • [29] Hyperbolic Insights With Knowledge Distillation for Cross-Domain Few-Shot Learning
    Yang, Xi
    Kong, Dechen
    Wang, Nannan
    Gao, Xinbo
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 1921 - 1933
  • [30] Incorporating geometry knowledge into an incremental learning structure for few-shot intent recognition
    Zhang, Xin
    Jiang, Miao
    Chen, Honghui
    Zheng, Jianming
    Pan, Zhiqiang
    KNOWLEDGE-BASED SYSTEMS, 2022, 251