Mitigate forgetting in few-shot class-incremental learning using different image views

被引:3
作者
Mazumder, Pratik [1 ]
Singh, Pravendra [2 ]
机构
[1] Indian Inst Technol Jodhpur, Jheepasani, India
[2] Indian Inst Technol Roorkee, Roorkee, India
关键词
Incremental learning; Few -shot learning; Catastrophic forgetting; Image classification;
D O I
10.1016/j.neunet.2023.06.043
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the few-shot class incremental learning (FSCIL) setting, new classes with few training examples become available incrementally, and deep learning models suffer from catastrophic forgetting of the previous classes when trained on new classes. Data augmentation techniques are generally used to increase the training data and improve the model performance. In this work, we demonstrate that differently augmented views of the same image obtained by applying data augmentations may not necessarily activate the same set of neurons in the model. Therefore, the information gained by a model regarding a class, when trained using data augmentation, may not necessarily be stored in the same set of neurons in the model. Consequently, during incremental training, even if some of the model weights that store the previously seen class information for a particular view get overwritten, the information of the previous classes for the other views may still remain intact in the other model weights. Therefore, the impact of catastrophic forgetting on the model predictions is different for different data augmentations used during training. Based on this, we present an Augmentation-based Prediction Rectification (APR) approach to reduce the impact of catastrophic forgetting in the FSCIL setting. APR can also augment other FSCIL approaches and significantly improve their performance. We also propose a novel feature synthesis module (FSM) for synthesizing features relevant to the previously seen classes without requiring training data from these classes. FSM outperforms other generative approaches in this setting. We experimentally show that our approach outperforms other methods on benchmark datasets. & COPY; 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:999 / 1009
页数:11
相关论文
共 65 条
[1]  
Achituve Idan, 2021, P MACHINE LEARNING R, V139
[2]   Task-Free Continual Learning [J].
Aljundi, Rahaf ;
Kelchtermans, Klaas ;
Tuytelaars, Tinne .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11246-11255
[3]   Class Incremental Learning With Few-Shots Based on Linear Programming for Hyperspectral Image Classification [J].
Bai, Jing ;
Yuan, Anran ;
Xiao, Zhu ;
Zhou, Huaji ;
Wang, Dingchen ;
Jiang, Hongbo ;
Jiao, Licheng .
IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (06) :5474-5485
[4]   A comprehensive study of class incremental learning algorithms for visual tasks [J].
Belouadah, Eden ;
Popescu, Adrian ;
Kanellos, Ioannis .
NEURAL NETWORKS, 2021, 135 :38-54
[5]   End-to-End Incremental Learning [J].
Castro, Francisco M. ;
Marin-Jimenez, Manuel J. ;
Guil, Nicolas ;
Schmid, Cordelia ;
Alahari, Karteek .
COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 :241-257
[6]  
Chaudhry A., 2019, P INT C LEARN REPR
[7]  
Chen K., 2021, INT C LEARNING REPRE
[8]  
Chen K., 2020, INT C LEARNING REPRE
[9]   Synthesized Feature based Few-Shot Class-Incremental Learning on a Mixture of Subspaces [J].
Cheraghian, Ali ;
Rahman, Shafin ;
Ramasinghe, Sameera ;
Fang, Pengfei ;
Simon, Christian ;
Petersson, Lars ;
Harandi, Mehrtash .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :8641-8650
[10]   Semantic-aware Knowledge Distillation for Few-Shot Class-Incremental Learning [J].
Cheraghian, Ali ;
Rahman, Shafin ;
Fang, Pengfei ;
Roy, Soumava Kumar ;
Petersson, Lars ;
Harandi, Mehrtash .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :2534-2543