Generative Models from the perspective of Continual Learning

被引:16
作者
Lesort, Timothee [1 ,2 ,3 ]
Caselles-Dupre, Hugo [1 ,2 ,4 ]
Garcia-Ortiz, Michael [4 ]
Stoian, Andrei [3 ]
Filliat, David [1 ,2 ]
机构
[1] ENSTA ParisTech, Flowers Team, Palaiseau, France
[2] INRIA, Paris, France
[3] Thales, Theresis Laboratory, La Defens, France
[4] Softbank Robot Europe, Paris, France
来源
2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN) | 2019年
关键词
D O I
10.1109/ijcnn.2019.8851986
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Which generative model is the most suitable for Continual Learning? This paper aims at evaluating and comparing generative models on disjoint sequential image generation tasks. We investigate how several models learn and forget, considering various strategies: rehearsal, regularization, generative replay and fine-tuning. We used two quantitative metrics to estimate the generation quality and memory ability. We experiment with sequential tasks on three commonly used benchmarks for Continual Learning (MNIST, Fashion MNIST and CIFAR10). We found that among all models, the original GAN performs best and among Continual Learning strategies, generative replay outperforms all other methods. Even if we found satisfactory combinations on MNIST and Fashion MNIST, training generative models sequentially on CIFAR10 is particularly instable, and remains a challenge. Our code is available online(1).
引用
收藏
页数:8
相关论文
共 50 条
  • [31] Generative Models for Relief Perspective Architectures
    Baglioni, Leonardo
    Fallavollita, Federico
    NEXUS NETWORK JOURNAL, 2021, 23 (04) : 879 - 898
  • [32] Learning Deep Generative Models
    Salakhutdinov, Ruslan
    ANNUAL REVIEW OF STATISTICS AND ITS APPLICATION, VOL 2, 2015, 2 : 361 - 385
  • [33] THE EFFECTIVENESS OF WORLD MODELS FOR CONTINUAL REINFORCEMENT LEARNING
    Kessler, Samuel
    Ostaszewski, Mateusz
    Bortkiewicz, Michal
    Zarski, Mateusz
    Wolczyk, Maciej
    Parker-Holder, Jack
    Roberts, Stephen J.
    Milos, Piotr
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 184 - 204
  • [34] ENERGY- BASED MODELS FOR CONTINUAL LEARNING
    Li, Shuang
    Du, Yilun
    van de Ven, Gido M.
    Mordatch, Igor
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 199, 2022, 199
  • [35] Exploring Continual Learning for Code Generation Models
    Yadav, Prateek
    Sun, Qing
    Ding, Hantian
    Li, Xiaopeng
    Zhang, Dejiao
    Tan, Ming
    Ma, Xiaofei
    Bhatia, Parminder
    Nallapati, Ramesh
    Ramanathan, Murali Krishna
    Bansal, Mohit
    Xiang, Bing
    61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 782 - 792
  • [36] Spiking Generative Adversarial Networks With a Neural Network Discriminator: Local Training, Bayesian Models, and Continual Meta-Learning
    Rosenfeld, Bleema
    Simeone, Osvaldo
    Rajendran, Bipin
    IEEE TRANSACTIONS ON COMPUTERS, 2022, 71 (11) : 2778 - 2791
  • [37] Revisiting Neural Networks for Continual Learning: An Architectural Perspective
    Lu, Aojun
    Feng, Tao
    Yuan, Hangjie
    Song, Xiaotian
    Sun, Yanan
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 4651 - 4659
  • [38] Continual Variational Autoencoder via Continual Generative Knowledge Distillation
    Ye, Fei
    Bors, Adrian G.
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10918 - 10926
  • [39] Turbulence scaling from deep learning diffusion generative models
    Whittaker, Tim
    Janik, Romuald A.
    Oz, Yaron
    JOURNAL OF COMPUTATIONAL PHYSICS, 2024, 514
  • [40] Learning Generative Models for Climbing Aircraft from Radar Data
    Pepper, Nick
    Thomas, Marc
    JOURNAL OF AEROSPACE INFORMATION SYSTEMS, 2024, 21 (06): : 474 - 481