Logarithmic Continual Learning

被引:1
作者
Masarczyk, Wojciech [1 ]
Wawrzynski, Pawel [1 ]
Marczak, Daniel [1 ]
Deja, Kamil [1 ]
Trzcinski, Tomasz [1 ,2 ,3 ]
机构
[1] Warsaw Univ Technol, Fac Elect & Informat Technol, PL-00661 Warsaw, Poland
[2] Jagiellonian Univ, Fac Math & Comp Sci, PL-31007 Krakow, Poland
[3] Tooploox, PL-53601 Wroclaw, Poland
关键词
Deep learning; Data models; Memory management; Computational modeling; Learning systems; Image reconstruction; Adaptation models; Continual learning; deep learning; incremental learning; rehearsal methods;
D O I
10.1109/ACCESS.2022.3218907
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We introduce a neural network architecture that logarithmically reduces the number of self-rehearsal steps in the generative rehearsal of continually learned models. In continual learning (CL), training samples come in subsequent tasks, and the trained model can access only a current task. Contemporary CL methods employ generative models to replay previous samples and train them recursively with a combination of current and regenerated past data. This recurrence leads to superfluous computations as the same past samples are regenerated after each task, and the reconstruction quality successively degrades. In this work, we address these limitations and propose a new generative rehearsal architecture that requires, at most, a logarithmic number of retraining sessions for each sample. Our approach leverages the allocation of past data in a set of generative models such that most of them do not require retraining after a task. The experimental evaluation of our logarithmic continual learning approach shows the superiority of our method with respect to the state-of-the-art generative rehearsal methods.
引用
收藏
页码:117001 / 117010
页数:10
相关论文
共 38 条
[1]  
Rusu AA, 2016, Arxiv, DOI [arXiv:1606.04671, 10.48550/arXiv.1606.04671, DOI 10.43550/ARXIV:1606.04671, DOI 10.48550/ARXIV.1606.04671]
[2]  
Aljundi R, 2019, ADV NEUR IN, V32
[3]  
Aljundi Rahaf, 2019, arXiv
[4]  
Caccia Lucas, 2020, INT C MACH LEARN, P1240
[5]  
Carreira-Perpiñán MA, 2015, PROC CVPR IEEE, P557, DOI 10.1109/CVPR.2015.7298654
[6]  
Deja K., 2021, PROC INT JOINT C NEU, P1
[7]   Continual Learning for Anomaly Detection in Surveillance Videos [J].
Doshi, Keval ;
Yilmaz, Yasin .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, :1025-1034
[8]   Catastrophic forgetting in connectionist networks [J].
French, RM .
TRENDS IN COGNITIVE SCIENCES, 1999, 3 (04) :128-135
[9]  
Golkar S., 2019, ARXIV190304476
[10]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672