Adaptive Progressive Continual Learning

被引:18
作者
Xu, Ju [1 ]
Ma, Jin [1 ]
Gao, Xuesong [2 ,3 ,4 ]
Zhu, Zhanxing [5 ]
机构
[1] Peking Univ, Ctr Data Sci, Beijing 100871, Peoples R China
[2] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300072, Peoples R China
[3] Hisense Co Ltd, State Key Lab Digital Multimedia Technol, Qingdao 266071, Shandong, Peoples R China
[4] Shandong Univ, Sch Informat Sci & Engn, Qingdao 266510, Shandong, Peoples R China
[5] Beijing Inst Big Data Res, Beijing 100124, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Optimization; Bayes methods; Training; Reinforcement learning; Knowledge engineering; Complexity theory; Machine learning; adaptive progressive network framework; continual learning; Bayesian optimization; reinforcement learning; neural networks;
D O I
10.1109/TPAMI.2021.3095064
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continual learning paradigm learns from a continuous stream of tasks in an incremental manner and aims to overcome the notorious issue: the catastrophic forgetting. In this work, we propose a new adaptive progressive network framework including two models for continual learning: Reinforced Continual Learning (RCL) and Bayesian Optimized Continual Learning with Attention mechanism (BOCL) to solve this fundamental issue. The core idea of this framework is to dynamically and adaptively expand the neural network structure upon the arrival of new tasks. RCL and BOCL employ reinforcement learning and Bayesian optimization to achieve it, respectively. An outstanding advantage of our proposed framework is that it will not forget the knowledge that has been learned through adaptively controlling the architecture. We propose effective ways of employing the learned knowledge in the two methods to control the size of the network. RCL employs previous knowledge directly while BOCL selectively utilizes previous knowledge (e.g., feature maps of previous tasks) via attention mechanism. The experiments on variants of MNIST, CIFAR-100 and Sequence of 5-Datasets demonstrate that our methods outperform the state-of-the-art in preventing catastrophic forgetting and fitting new tasks better under the same or less computing resource.
引用
收藏
页码:6715 / 6728
页数:14
相关论文
共 50 条
[21]   Adaptive Online Domain Incremental Continual Learning [J].
Gunasekara, Nuwan ;
Gomes, Heitor ;
Bifet, Albert ;
Pfahringer, Bernhard .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT I, 2022, 13529 :491-502
[22]   An Adaptive Continual Learning Method for Nonstationary Industrial Time Series Prediction [J].
Wu, Mengqing ;
Zhou, Xiaofeng ;
Li, Shuai ;
Shi, Haibo .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2025, 21 (02) :1160-1169
[23]   VIPeR: Visual Incremental Place Recognition With Adaptive Mining and Continual Learning [J].
Ming, Yuhang ;
Xu, Minyang ;
Yang, Xingrui ;
Ye, Weicai ;
Wang, Weihan ;
Peng, Yong ;
Dai, Weichen ;
Kong, Wanzeng .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (03) :3038-3045
[24]   Continual Learning With Structured Inheritance for Semantic Segmentation in Aerial Imagery [J].
Feng, Yingchao ;
Sun, Xian ;
Diao, Wenhui ;
Li, Jihao ;
Gao, Xin ;
Fu, Kun .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
[25]   SPACE: Structured Compression and Sharing of Representational Space for Continual Learning [J].
Saha, Gobinda ;
Garg, Isha ;
Ankit, Aayush ;
Roy, Kaushik .
IEEE ACCESS, 2021, 9 :150480-150494
[26]   Neural Agents with Continual Learning Capacities [J].
Zhinin-Vera, Luis ;
Pretel, Elena ;
Moya, Alejandro ;
Jimenez-Ruescas, Javier ;
Astudillo, Jaime .
INFORMATION AND COMMUNICATION TECHNOLOGIES, TICEC 2024, 2025, 2273 :145-159
[27]   Continual Learning of Generative Models With Limited Data: From Wasserstein-1 Barycenter to Adaptive Coalescence [J].
Dedeoglu, Mehmet ;
Lin, Sen ;
Zhang, Zhaofeng ;
Zhang, Junshan .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (09) :12042-12056
[28]   Unsupervised Continual Learning in Streaming Environments [J].
Ashfahani, Andri ;
Pratama, Mahardhika .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (12) :9992-10003
[29]   Block Contextual MDPs for Continual Learning [J].
Sodhani, Shagun ;
Meier, Franziska ;
Pineau, Joelle ;
Zhang, Amy .
LEARNING FOR DYNAMICS AND CONTROL CONFERENCE, VOL 168, 2022, 168
[30]   Continual Learning With Knowledge Distillation: A Survey [J].
Li, Songze ;
Su, Tonghua ;
Zhang, Xuyao ;
Wang, Zhongjie .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (06) :9798-9818