Continual compression model for online continual learning

被引:0
作者
Ye, Fei [1 ]
Bors, Adrian G. [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu, Peoples R China
[2] Univ York, Dept Comp Sci, York YO10 5GH, England
关键词
Continual learning; Dynamic expansion model; Task-Free Continual Learning; Component pruning;
D O I
10.1016/j.asoc.2024.112427
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Task-Free Continual Learning (TFCL) presents a notably demanding but realistic ongoing learning concept, aiming to address catastrophic forgetting in sequential learning systems. In this paper, we tackle catastrophic forgetting by introducing an innovative dynamic expansion framework designed to adaptively enhance the model's capacity for novel data learning while also remembering the information learnt in the past, by using a minimal-size processing architecture. Our proposed framework incorporates three key mechanisms to mitigate model' forgetting: (1) by employing a Maximum Mean Discrepancy (MMD)-based expansion mechanism that assesses the disparity between previously acquired knowledge and that from the new training data, serving as a signal for the model's architecture expansion; (2) a component discarding mechanism that eliminates components characterized by redundant information, thereby optimizing the model size while fostering knowledge diversity; (3) a novel training sample selection strategy that leads to the diversity of the training data for each task. We conduct a series of TFCL experiments that demonstrate the superiority of the proposed framework over all baselines while utilizing fewer components than alternative dynamic expansion models. The results on the Split Mini ImageNet dataset, after splitting the original dataset into multiple tasks, are improved by more than 2% when compared to the closest baseline.
引用
收藏
页数:14
相关论文
共 56 条
[1]  
Rusu AA, 2016, Arxiv, DOI [arXiv:1606.04671, DOI 10.48550/ARXIV.1606.04671, DOI 10.43550/ARXIV:1606.04671]
[2]  
Aljundi R, 2019, ADV NEUR IN, V32
[3]  
Aljundi R, 2019, ADV NEUR IN, V32
[4]   Task-Free Continual Learning [J].
Aljundi, Rahaf ;
Kelchtermans, Klaas ;
Tuytelaars, Tinne .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :11246-11255
[5]   Expert Gate: Lifelong Learning with a Network of Experts [J].
Aljundi, Rahaf ;
Chakravarty, Punarjay ;
Tuytelaars, Tinne .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :7120-7129
[6]   Rainbow Memory: Continual Learning with a Memory of Diverse Samples [J].
Bang, Jihwan ;
Kim, Heesu ;
Yoo, YoungJoon ;
Ha, Jung-Woo ;
Choi, Jonghyun .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :8214-8223
[7]  
Buzzega P, 2020, Advances in neural information processing systems, V33, P15920
[8]  
Chaudhry A, 2019, arXiv, DOI 10.48550/arXiv.1902.10486
[9]   Max-Margin Deep Diverse Latent Dirichlet Allocation With Continual Learning [J].
Chen, Wenchao ;
Chen, Bo ;
Liu, Yingqi ;
Cao, Xuefei ;
Zhao, Qianru ;
Zhang, Hao ;
Tian, Long .
IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (07) :5639-5653
[10]  
Cortes C, 2017, PR MACH LEARN RES, V70