Adapt & Align: Continual Learning with Generative Models' Latent Space Alignment

被引:0
作者
Deja, Kamil [1 ,2 ]
Cywinski, Bartosz [1 ,2 ]
Rybarczyk, Jan [1 ]
Trzcinski, Tomasz [1 ,2 ,3 ]
机构
[1] Warsaw Univ Technol, pl Politech 1, PL-00661 Warsaw, Poland
[2] Ideas Res Inst, ul Krolewska 27, PL-00060 Warsaw, Poland
[3] Tooploox, Teczowa 7, PL-53601 Wroclaw, Poland
关键词
Continual learning; Generative modeling; Vae; Gan;
D O I
10.1016/j.neucom.2025.130748
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Motivation: Neural networks suffer from abrupt loss in performance when retrained with additional data from different distributions. At the same time, training with additional data without access to the previous examples rarely improves the model's performance. Methods: We propose Adapt & Align, a novel continual learning framework that leverages generative models to align their latent representations across tasks. The approach is divided into two phases: focil Traffing Train I gencrative model (eg, a Vaflational Autbencoder TVA) ora Gellertive Adversarial Network (GAN)) on the current task to capture task-specific features. Global Training: Use a translator network to map these task-specific latent representations into a unified global latent space, thereby facilitating both forward and backward knowledge transfer. Results: Experiments on benchmark datasets (c.g., MNIST, Omniglot, CIFAR, CelebA) as well as real-world application for particle simulation at CERN demonstrate that Adapt & Align mitigates catastrophic forgetting and improves generation quality as indicated by metrics such as Fr & eacute;chet Inception Distance (FID), distribution precision and recall, or accuracy for the downstream classification task. Ablation studies confirm the critical role of cach component.
引用
收藏
页数:17
相关论文
共 74 条
[1]  
Achille A, 2018, ADV NEUR IN, V31
[2]  
Alemohammad S, 2023, Arxiv, DOI [arXiv:2307.01850, DOI 10.48550/ARXIV.2307.01850]
[3]   Memory Aware Synapses: Learning What (not) to Forget [J].
Aljundi, Rahaf ;
Babiloni, Francesca ;
Elhoseiny, Mohamed ;
Rohrbach, Marcus ;
Tuytelaars, Tinne .
COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 :144-161
[4]  
Arjovsky M, 2017, PR MACH LEARN RES, V70
[5]  
Baranchuk Dmitry, 2021, INT C LEARN REPR
[6]  
Binkowski M., 2018, INT C LEARN REPR ICL
[7]  
Brock A., 2018, INT C LEARN REPR, DOI DOI 10.1145/3205873.3205877
[8]  
Deja K, 2022, PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, P2902
[9]  
Deja K, 2023, Arxiv, DOI [arXiv:2301.13622, 10.48550/arXiv.2301.13622, DOI 10.48550/ARXIV.2301.13622]
[10]   End-to-End Sinkhorn Autoencoder With Noise Generator [J].
Deja, Kamil ;
Dubinski, Jan ;
Nowak, Piotr ;
Wenzel, Sandro ;
Spurek, Przemyslaw ;
Trzcinski, Tomasz .
IEEE ACCESS, 2021, 9 :7211-7219