Continual compression model for online continual learning

被引:0
|
作者
Ye, Fei [1 ]
Bors, Adrian G. [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu, Peoples R China
[2] Univ York, Dept Comp Sci, York YO10 5GH, England
关键词
Continual learning; Dynamic expansion model; Task-Free Continual Learning; Component pruning;
D O I
10.1016/j.asoc.2024.112427
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Task-Free Continual Learning (TFCL) presents a notably demanding but realistic ongoing learning concept, aiming to address catastrophic forgetting in sequential learning systems. In this paper, we tackle catastrophic forgetting by introducing an innovative dynamic expansion framework designed to adaptively enhance the model's capacity for novel data learning while also remembering the information learnt in the past, by using a minimal-size processing architecture. Our proposed framework incorporates three key mechanisms to mitigate model' forgetting: (1) by employing a Maximum Mean Discrepancy (MMD)-based expansion mechanism that assesses the disparity between previously acquired knowledge and that from the new training data, serving as a signal for the model's architecture expansion; (2) a component discarding mechanism that eliminates components characterized by redundant information, thereby optimizing the model size while fostering knowledge diversity; (3) a novel training sample selection strategy that leads to the diversity of the training data for each task. We conduct a series of TFCL experiments that demonstrate the superiority of the proposed framework over all baselines while utilizing fewer components than alternative dynamic expansion models. The results on the Split Mini ImageNet dataset, after splitting the original dataset into multiple tasks, are improved by more than 2% when compared to the closest baseline.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Online continual learning through unsupervised mutual information maximization
    Hihn, Heinke
    Braun, Daniel A.
    NEUROCOMPUTING, 2024, 578
  • [22] Online Continual Learning in Acoustic Scene Classification: An Empirical Study
    Ha, Donghee
    Kim, Mooseop
    Jeong, Chi Yoon
    SENSORS, 2023, 23 (15)
  • [23] Open-world continual learning: Unifying novelty detection and continual learning
    Kim, Gyuhak
    Xiao, Changnan
    Konishi, Tatsuya
    Ke, Zixuan
    Liu, Bing
    ARTIFICIAL INTELLIGENCE, 2025, 338
  • [24] Continual Variational Autoencoder Learning via Online Cooperative Memorization
    Ye, Fei
    Bors, Adrian G.
    COMPUTER VISION, ECCV 2022, PT XXIII, 2022, 13683 : 531 - 549
  • [25] Continual Representation Learning for Images with Variational Continual Auto-Encoder
    Jeon, Ik Hwan
    Shin, Soo Young
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 367 - 373
  • [26] Continual variational dropout: a view of auxiliary local variables in continual learning
    Nam Le Hai
    Trang Nguyen
    Linh Ngo Van
    Thien Huu Nguyen
    Khoat Than
    Machine Learning, 2024, 113 : 281 - 323
  • [27] ConCS: A Continual Classifier System for Continual Learning of Multiple Boolean Problems
    Nguyen, Trung B.
    Browne, Will N.
    Zhang, Mengjie
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2023, 27 (04) : 1057 - 1071
  • [28] Continual variational dropout: a view of auxiliary local variables in continual learning
    Hai, Nam Le
    Nguyen, Trang
    Van, Linh Ngo
    Nguyen, Thien Huu
    Than, Khoat
    MACHINE LEARNING, 2024, 113 (01) : 281 - 323
  • [29] Beyond Prompt Learning: Continual Adapter for Efficient Rehearsal-Free Continual Learning
    Gao, Xinyuan
    Dong, Songlin
    He, Yuhang
    Wang, Qiang
    Gong, Yihong
    COMPUTER VISION - ECCV 2024, PT LXXXV, 2025, 15143 : 89 - 106
  • [30] MAGMAX: Leveraging Model Merging for Seamless Continual Learning
    Marczak, Daniel
    Twardowski, Bartiomiej
    Trzcinski, Tomasz
    Cygert, Sebastian
    COMPUTER VISION - ECCV 2024, PT LXXXV, 2025, 15143 : 379 - 395