Continual compression model for online continual learning

被引:0
|
作者
Ye, Fei [1 ]
Bors, Adrian G. [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu, Peoples R China
[2] Univ York, Dept Comp Sci, York YO10 5GH, England
关键词
Continual learning; Dynamic expansion model; Task-Free Continual Learning; Component pruning;
D O I
10.1016/j.asoc.2024.112427
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Task-Free Continual Learning (TFCL) presents a notably demanding but realistic ongoing learning concept, aiming to address catastrophic forgetting in sequential learning systems. In this paper, we tackle catastrophic forgetting by introducing an innovative dynamic expansion framework designed to adaptively enhance the model's capacity for novel data learning while also remembering the information learnt in the past, by using a minimal-size processing architecture. Our proposed framework incorporates three key mechanisms to mitigate model' forgetting: (1) by employing a Maximum Mean Discrepancy (MMD)-based expansion mechanism that assesses the disparity between previously acquired knowledge and that from the new training data, serving as a signal for the model's architecture expansion; (2) a component discarding mechanism that eliminates components characterized by redundant information, thereby optimizing the model size while fostering knowledge diversity; (3) a novel training sample selection strategy that leads to the diversity of the training data for each task. We conduct a series of TFCL experiments that demonstrate the superiority of the proposed framework over all baselines while utilizing fewer components than alternative dynamic expansion models. The results on the Split Mini ImageNet dataset, after splitting the original dataset into multiple tasks, are improved by more than 2% when compared to the closest baseline.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Memory Bounds for Continual Learning
    Chen, Xi
    Papadimitriou, Christos
    Peng, Binghui
    2022 IEEE 63RD ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS), 2022, : 519 - 530
  • [32] Reinforced Continual Learning for Graphs
    Rakaraddi, Appan
    Kei, Lam Siew
    Pratama, Mahardhika
    de Carvalho, Marcus
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 1666 - 1674
  • [33] Continual learning with selective nets
    Luu, Hai Tung
    Szemenyei, Marton
    APPLIED INTELLIGENCE, 2025, 55 (07)
  • [34] Continual Information Cascade Learning
    Zhou, Fan
    Jing, Xin
    Xu, Xovee
    Zhong, Ting
    Trajcevski, Goce
    Wu, Jin
    2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [35] Advances and Trends of Continual Learning
    Li, Wenbin
    Xiong, Yakun
    Fan, Zhichen
    Deng, Bo
    Cao, Fuyuan
    Gao, Yang
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2024, 61 (06): : 1476 - 1496
  • [36] Continual Learning, Fast and Slow
    Pham, Quang
    Liu, Chenghao
    Hoi, Steven C. H.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (01) : 134 - 149
  • [37] Adaptive Progressive Continual Learning
    Xu, Ju
    Ma, Jin
    Gao, Xuesong
    Zhu, Zhanxing
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (10) : 6715 - 6728
  • [38] Continual learning in the presence of repetition
    Hemati, Hamed
    Pellegrini, Lorenzo
    Duan, Xiaotian
    Zhao, Zixuan
    Xia, Fangfang
    Masana, Marc
    Tscheschner, Benedikt
    Veas, Eduardo
    Zheng, Yuxiang
    Zhao, Shiji
    Li, Shao-Yuan
    Huang, Sheng-Jun
    Lomonaco, Vincenzo
    van de Ven, Gido M.
    NEURAL NETWORKS, 2025, 183
  • [39] Online Continual Learning of End-to-End Speech Recognition Models
    Yang, Muqiao
    Lane, Ian
    Watanabe, Shinji
    INTERSPEECH 2022, 2022, : 2668 - 2672
  • [40] Toward an Online Continual Learning Architecture for Intrusion Detection of Video Surveillance
    Kwon, Beom
    Kim, Taewan
    IEEE ACCESS, 2022, 10 : 89732 - 89744