On the Beneficial Effects of Reinjections for Continual Learning

被引:0
作者
Solinas M. [1 ]
Reyboz M. [1 ]
Rousset S. [1 ,2 ]
Galliere J. [1 ]
Mainsant M. [1 ]
Bourrier Y. [1 ,2 ]
Molnos A. [1 ]
Mermillod M. [1 ,2 ]
机构
[1] Univ. Grenoble Alpes, CEA, LIST, Grenoble
[2] LPNC, Univ Grenoble Alpes, Univ Savoie Mont Blanc, Grenoble
关键词
Continual learning; Incremental learning; Lifelong learning; Pseudo-rehearsal; Rehearsal; Sequential learning;
D O I
10.1007/s42979-022-01392-7
中图分类号
学科分类号
摘要
Deep learning delivers remarkable results in a wide range of applications, but artificial neural networks still suffer from catastrophic forgetting of old knowledge as new knowledge is learned. Rehearsal methods overcome catastrophic forgetting by replaying an amount of previously learned data stored in dedicated memory buffers. Alternatively, pseudo-rehearsal methods generate pseudo-samples to emulate previously learned data, alleviating the need for dedicated buffers. First, we show that it is possible to alleviate catastrophic forgetting with a pseudo-rehearsal method without employing memory buffers or generative models. We propose a hybrid architecture similar to that of an autoencoder with additional neurons to classify the input. This architecture preserves specific properties of autoencoders by allowing the generation of pseudo-samples through reinjections (i.e. iterative sampling) from random noise. The generated pseudo-samples are then interwoven with the new examples to acquire new knowledge without forgetting the previous ones. Second, we combine the two methods (rehearsal and pseudo-rehearsal) in the hybrid architecture. Examples stored in small memory buffers are employed as seeds instead of noise to improve the process of generating pseudo-samples and retrieving previously learned knowledge. We demonstrate that reinjections are suitable for rehearsal and pseudo-rehearsal approaches and show state-of-the-art results on rehearsal methods for small buffer sizes. We evaluate our method extensively on MNIST, CIFAR-10 and CIFAR-100 image classification datasets. © 2022, The Author(s), under exclusive licence to Springer Nature Singapore Pte Ltd.
引用
收藏
相关论文
共 100 条
[1]  
Kemker R., McClure M., Abitino A., Hayes T., Kanan C., Measuring catastrophic forgetting in neural networks, Proceedings of the AAAI Conference on Artificial Intelligence, 32, (2018)
[2]  
Parisi G.I., Kemker R., Part J.L., Kanan C., Wermter S., Continual lifelong learning with neural networks: a review, Neural Netw, 113, pp. 54-71, (2019)
[3]  
de Lange M., Aljundi R., Masana M., Parisot S., Jia X., Leonardis A., Slabaugh G., Tuytelaars T., A continual learning survey: Defying forgetting in classification tasks, (2019)
[4]  
Masana M., Liu X., Twardowski B., Menta M., Bagdanov A.D., van De Weijer J., Class-Incremental Learning: Survey and Performance Evaluation, (2020)
[5]  
Mai Z., Li R., Jeong J., Quispe D., Kim H., Sanner. S., Online continual learning in image classification: An empirical survey, 2021. Arxiv Preprint Arxiv
[6]  
McCloskey M., Cohen N.J., Catastrophic interference in connectionist networks: the sequential learning problem, Psychol Learn Motiv, 24, pp. 109-165, (1989)
[7]  
Akers K.G., Martinez-Canabal A., Restivo L., Yiu A.P., De Cristofaro A., Hsiang H.-L.L., Wheeler A.L., Guskjolen A., Niibori Y., Shoji H., Et al., Hippocampal neurogenesis regulates forgetting during adulthood and infancy, Science, 344, 6184, pp. 598-602, (2014)
[8]  
Yang G., Lai C.S.W., Cichon J., Ma L., Li W., Gan W.-B., Sleep promotes branch-specific formation of dendritic spines after learning, Science, 344, 6188, pp. 1173-1178, (2014)
[9]  
Barry D.N., Maguire E.A., Remote memory and the hippocampus: a constructive critique, Trends Cogn Sci, 23, 2, pp. 128-142, (2019)
[10]  
Fernando C., Banarse D., Blundell C., Zwols Y., Ha D., Rusu A.A., Pritzel A., Wierstra D., Pathnet: Evolution channels gradient descent in super neural networks, . Arxiv Preprint Arxiv, (2017)