Relational Experience Replay: Continual Learning by Adaptively Tuning Task-Wise Relationship

被引:1
作者
Wang, Quanziang [1 ]
Wang, Renzhen [1 ]
Li, Yuexiang [2 ]
Wei, Dong [2 ]
Wang, Hong [2 ]
Ma, Kai [2 ]
Zheng, Yefeng [2 ]
Meng, Deyu [1 ,3 ,4 ]
机构
[1] Xi An Jiao Tong Univ, Sch Math & Stat, Xian 710049, Peoples R China
[2] Tencent Jarvis Lab, Shenzhen 518052, Peoples R China
[3] Xi An Jiao Tong Univ, Key Lab Intelligent Networks & Network Secur, Minist Educ, Xian 710049, Peoples R China
[4] Macau Univ Sci & Technol, Macao Inst Syst Engn, Macau, Peoples R China
基金
国家重点研发计划;
关键词
Task analysis; Training; Optimization; Adaptation models; Data models; Tuning; Noise measurement; Continual learning; stability-plasticity dilemma; bi-level optimization;
D O I
10.1109/TMM.2024.3397048
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Continual learning is a promising machine learning paradigm to learn new tasks while retaining previously learned knowledge over streaming training data. Till now, rehearsal-based methods, keeping a small part of data from old tasks as a memory buffer, have shown good performance in mitigating catastrophic forgetting for previously learned knowledge. However, most of these methods typically treat each new task equally, which may not adequately consider the relationship or similarity between old and new tasks. Furthermore, these methods commonly neglect sample importance in the continual training process and result in sub-optimal performance on certain tasks. To address this challenging problem, we propose Relational Experience Replay (RER), a bi-level learning framework, to adaptively tune task-wise relationships and sample importance within each task to achieve a better 'stability' and 'plasticity' trade-off. As such, the proposed method is capable of accumulating new knowledge while consolidating previously learned old knowledge during continual learning. Extensive experiments conducted on three benchmark image datasets (CIFAR-10, CIFAR-100, and Tiny ImageNet) and two text datasets (20News and DBpedia) show that the proposed method can consistently improve the performance of all baselines and surpass current state-of-the-art methods.
引用
收藏
页码:9683 / 9698
页数:16
相关论文
共 60 条
[1]  
Rusu AA, 2016, Arxiv, DOI arXiv:1606.04671
[2]   Conditional Channel Gated Networks for Task-Aware Continual Learning [J].
Abati, Davide ;
Tomczak, Jakub ;
Blankevoort, Tijmen ;
Calderara, Simone ;
Cucchiara, Rita ;
Bejnordi, Babak Ehteshami .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :3930-3939
[3]   SS-IL: Separated Softmax for Incremental Learning [J].
Ahn, Hongjoon ;
Kwak, Jihwan ;
Lim, Subin ;
Bang, Hyeonsu ;
Kim, Hyojun ;
Moon, Taesup .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :824-833
[4]  
Aljundi R, 2019, ADV NEUR IN, V32
[5]  
[Anonymous], 2008, 20 newsgroups
[6]   Rainbow Memory: Continual Learning with a Memory of Diverse Samples [J].
Bang, Jihwan ;
Kim, Heesu ;
Yoo, YoungJoon ;
Ha, Jung-Woo ;
Choi, Jonghyun .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :8214-8223
[7]   DBpedia - A crystallization point for the Web of Data [J].
Bizer, Christian ;
Lehmann, Jens ;
Kobilarov, Georgi ;
Auer, Soeren ;
Becker, Christian ;
Cyganiak, Richard ;
Hellmann, Sebastian .
JOURNAL OF WEB SEMANTICS, 2009, 7 (03) :154-165
[8]  
Bochkovskiy A, 2020, Arxiv, DOI arXiv:2004.10934
[9]  
Buzzega P., 2020, ADV NEURAL INFORM PR, V33, P15920
[10]  
Caccia L., 2022, P INT C LEARN REPR