Continual learning with invertible generative models

被引:3
作者
Pomponi, Jary [1 ]
Scardapane, Simone [1 ]
Uncini, Aurelio [1 ]
机构
[1] Sapienza Univ Rome, Dept Informat Engn Elect & Telecommun DIET, Rome, Italy
关键词
Machine learning; Continual learning; Normalizing flow; Catastrophic forgetting; NEURAL-NETWORKS;
D O I
10.1016/j.neunet.2023.05.020
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Catastrophic forgetting (CF) happens whenever a neural network overwrites past knowledge while being trained on new tasks. Common techniques to handle CF include regularization of the weights (using, e.g., their importance on past tasks), and rehearsal strategies, where the network is constantly re-trained on past data. Generative models have also been applied for the latter, in order to have endless sources of data. In this paper, we propose a novel method that combines the strengths of regularization and generative-based rehearsal approaches. Our generative model consists of a normal-izing flow (NF), a probabilistic and invertible neural network, trained on the internal embeddings of the network. By keeping a single NF throughout the training process, we show that our memory overhead remains constant. In addition, exploiting the invertibility of the NF, we propose a simple approach to regularize the network's embeddings with respect to past tasks. We show that our method performs favorably with respect to state-of-the-art approaches in the literature, with bounded computational power and memory overheads.(c) 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:606 / 616
页数:11
相关论文
共 50 条
[21]   Progressive learning: A deep learning framework for continual learning [J].
Fayek, Haytham M. ;
Cavedon, Lawrence ;
Wu, Hong Ren .
NEURAL NETWORKS, 2020, 128 :345-357
[22]   Continual Learning From a Stream of APIs [J].
Yang, Enneng ;
Wang, Zhenyi ;
Shen, Li ;
Yin, Nan ;
Liu, Tongliang ;
Guo, Guibing ;
Wang, Xingwei ;
Tao, Dacheng .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) :11432-11445
[23]   Continual Learning With Knowledge Distillation: A Survey [J].
Li, Songze ;
Su, Tonghua ;
Zhang, Xuyao ;
Wang, Zhongjie .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (06) :9798-9818
[24]   Optimizing Foundation Models for Histopathology: A Continual Learning Approach to Cancer Detection [J].
Yaday, Ankur ;
Daescu, Ovidiu .
TRUSTWORTHY ARTIFICIAL INTELLIGENCE FOR HEALTHCARE, TAI4H 2024, 2024, 14812 :145-156
[25]   A comprehensive survey and analysis of generative models in machine learning [J].
Harshvardhan, G. M. ;
Gourisaria, Mahendra Kumar ;
Pandey, Manjusha ;
Rautaray, Siddharth Swarup .
COMPUTER SCIENCE REVIEW, 2020, 38
[26]   Predictive and generative machine learning models for photonic crystals [J].
Christensen, Thomas ;
Loh, Charlotte ;
Picek, Stjepan ;
Jing, Li ;
Fisher, Sophie ;
Ceperic, Vladimir ;
Joannopoulos, John D. ;
Soljacic, Marin ;
Jakobovic, Domagoj .
NANOPHOTONICS, 2020, 9 (13) :4183-4192
[27]   Dual Network Based Complementary Learning System for Continual Learning [J].
Kumari, Geeta ;
Song, Iickho .
2021 IEEE/CIC INTERNATIONAL CONFERENCE ON COMMUNICATIONS IN CHINA, ICCC WORKSHOPS, 2021, :112-117
[28]   Is Class-Incremental Enough for Continual Learning? [J].
Cossu, Andrea ;
Graffieti, Gabriele ;
Pellegrini, Lorenzo ;
Maltoni, Davide ;
Bacciu, Davide ;
Carta, Antonio ;
Lomonaco, Vincenzo .
FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2022, 5
[29]   Continual Learning for Object Classification: A Modular Approach [J].
Turner, Daniel ;
Cardoso, Pedro J. S. ;
Rodrigues, Joao M. F. .
UNIVERSAL ACCESS IN HUMAN-COMPUTER INTERACTION. ACCESS TO MEDIA, LEARNING AND ASSISTIVE ENVIRONMENTS, UAHCI 2021, PT II, 2021, 12769 :531-547
[30]   Continual Learning with Sparse Progressive Neural Networks [J].
Ergun, Esra ;
Toreyin, Behcet Ugur .
2020 28TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2020,