Adaptive Progressive Continual Learning

被引:13
|
作者
Xu, Ju [1 ]
Ma, Jin [1 ]
Gao, Xuesong [2 ,3 ,4 ]
Zhu, Zhanxing [5 ]
机构
[1] Peking Univ, Ctr Data Sci, Beijing 100871, Peoples R China
[2] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300072, Peoples R China
[3] Hisense Co Ltd, State Key Lab Digital Multimedia Technol, Qingdao 266071, Shandong, Peoples R China
[4] Shandong Univ, Sch Informat Sci & Engn, Qingdao 266510, Shandong, Peoples R China
[5] Beijing Inst Big Data Res, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Optimization; Bayes methods; Training; Reinforcement learning; Knowledge engineering; Complexity theory; Machine learning; adaptive progressive network framework; continual learning; Bayesian optimization; reinforcement learning; neural networks;
D O I
10.1109/TPAMI.2021.3095064
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continual learning paradigm learns from a continuous stream of tasks in an incremental manner and aims to overcome the notorious issue: the catastrophic forgetting. In this work, we propose a new adaptive progressive network framework including two models for continual learning: Reinforced Continual Learning (RCL) and Bayesian Optimized Continual Learning with Attention mechanism (BOCL) to solve this fundamental issue. The core idea of this framework is to dynamically and adaptively expand the neural network structure upon the arrival of new tasks. RCL and BOCL employ reinforcement learning and Bayesian optimization to achieve it, respectively. An outstanding advantage of our proposed framework is that it will not forget the knowledge that has been learned through adaptively controlling the architecture. We propose effective ways of employing the learned knowledge in the two methods to control the size of the network. RCL employs previous knowledge directly while BOCL selectively utilizes previous knowledge (e.g., feature maps of previous tasks) via attention mechanism. The experiments on variants of MNIST, CIFAR-100 and Sequence of 5-Datasets demonstrate that our methods outperform the state-of-the-art in preventing catastrophic forgetting and fitting new tasks better under the same or less computing resource.
引用
收藏
页码:6715 / 6728
页数:14
相关论文
共 50 条
  • [1] Progressive learning: A deep learning framework for continual learning
    Fayek, Haytham M.
    Cavedon, Lawrence
    Wu, Hong Ren
    NEURAL NETWORKS, 2020, 128 : 345 - 357
  • [2] Visual Tracking by Adaptive Continual Meta-Learning
    Choi, Janghoon
    Baik, Sungyong
    Choi, Myungsub
    Kwon, Junseok
    Lee, Kyoung Mu
    IEEE ACCESS, 2022, 10 : 9022 - 9035
  • [3] Multiagent Continual Coordination via Progressive Task Contextualization
    Yuan, Lei
    Li, Lihe
    Zhang, Ziqian
    Zhang, Fuxiang
    Guan, Cong
    Yu, Yang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 15
  • [4] Continual Learning of Knowledge Graph Embeddings
    Daruna, Angel
    Gupta, Mehul
    Sridharan, Mohan
    Chernova, Sonia
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (02) : 1128 - 1135
  • [5] Continual Learning with Sparse Progressive Neural Networks
    Ergun, Esra
    Toreyin, Behcet Ugur
    2020 28TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2020,
  • [6] A Continual Learning Survey: Defying Forgetting in Classification Tasks
    De Lange, Matthias
    Aljundi, Rahaf
    Masana, Marc
    Parisot, Sarah
    Jia, Xu
    Leonardis, Ales
    Slabaugh, Greg
    Tuytelaars, Tinne
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (07) : 3366 - 3385
  • [7] Efficient Architecture Search for Continual Learning
    Gao, Qiang
    Luo, Zhipeng
    Klabjan, Diego
    Zhang, Fengli
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 8555 - 8565
  • [8] Class-Incremental Continual Learning Into the eXtended DER-Verse
    Boschini, Matteo
    Bonicelli, Lorenzo
    Buzzega, Pietro
    Porrello, Angelo
    Calderara, Simone
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (05) : 5497 - 5512
  • [9] Sparse Progressive Neural Networks for Continual Learning
    Ergun, Esra
    Toreyin, Behcet Ugur
    ADVANCES IN COMPUTATIONAL COLLECTIVE INTELLIGENCE (ICCCI 2021), 2021, 1463 : 715 - 725
  • [10] PROGRESSIVE CONTINUAL LEARNING FOR SPOKEN KEYWORD SPOTTING
    Huang, Yizheng
    Hou, Nana
    Chen, Nancy F.
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7552 - 7556