A novel data-free continual learning method with contrastive reversion

被引:0
作者
Chu Wu
Runshan Xie
Shitong Wang
机构
[1] Jiangnan University,The School of AI & Computer Science
来源
International Journal of Machine Learning and Cybernetics | 2024年 / 15卷
关键词
Continual learning; Catastrophic forgetting; Data-free Knowledge distillation; Contrastive learning;
D O I
暂无
中图分类号
学科分类号
摘要
While continual learning has shown its impressive performance in addressing catastrophic forgetting of traditional neural networks and enabling them to learn multiple tasks continuously, it still requires a large amount of input data to train neural networks with satisfactory classification performance. Since collecting a large amount of training data is a time-consuming and expensive procedure, this study attempts to propose a novel data-free contrastive reversion method for continual learning (DFCRCL) to significantly reduce the number of training data for continual learning, while maintaining or even improving the classification performance of continual learning. In order to achieve such a goal, DFCRCL uses contrastive reversion to generate high-semantic pseudo samples from the previous task to guide the training of the current task. DFCRCL has three merits: (1) knowledge distillation from the previous task model to the current task model guarantees both the reduction of training data and the avoidance of catastrophic forgetting, and thus DFCRCL can effectively learn a sequence of tasks continuously (2) contrastive reversion enhances the semantic diversity of pseudo samples by learning the distinguishability between distinct pseudo samples in the feature space (3) contrastive reversion improves the performance of knowledge distillation in DFCRCL by enhancing the semantic diversity of the pseudo samples generated from the previous task model. Compared to six mainstream continual learning methods, the proposed DFCRCL achieves at least comparable or even better classification performance and stability in four benchmarking continual learning scenarios. In addition, the effectiveness of DFCRCL is demonstrated by ablation experiments.
引用
收藏
页码:505 / 518
页数:13
相关论文
共 61 条
  • [1] Zhang W(2020)Multimodel feature reinforcement framework using Moore-Penrose inverse for big data analysis IEEE Trans Neural Netw Learn Syst 32 5008-5021
  • [2] Wu QMJ(2022)Multimodal Moore-Penrose inverse-based recomputation framework for big data analysis IEEE Trans Neural Netw Learn Syst 3 128-135
  • [3] Yang Y(1999)Catastrophic forgetting in connectionist networks Trends Cogn Sci 97 285-207
  • [4] Zhang W(1990)Connectionist models of recognition memory: constraints imposed by learning and forgetting functions Psychol Rev 12 1-71
  • [5] Yang Y(2018)Lifelong machine learning Synth Lect Artif Intell Mach Learn 113 54-3385
  • [6] Wu QMJ(2019)Continual lifelong learning with neural networks: a review Neural Netw 44 3366-25
  • [7] French RM(2021)A continual learning survey: defying forgetting in classification tasks IEEE Trans Pattern Anal Mach Intell 10 12-1934
  • [8] Ratcliff R(2015)Learning in nonstationary environments: a survey IEEE Comput Intell Mag 33 1925-5542
  • [9] Chen Z(2021)Triple-memory networks: a brain-inspired method for continual learning IEEE Trans Neural Netw Learn Syst 2017 5533-3526
  • [10] Liu B(2017)iCaRL: incremental classifier and representation learning IEEE Conf Comput Vis Pattern Recogn (CVPR) 114 3521-2947