Continual Learning Using Bayesian Neural Networks

被引:24
|
作者
Li, Honglin [1 ,2 ]
Barnaghi, Payam [1 ,2 ,3 ]
Enshaeifare, Shirin [1 ,2 ]
Ganz, Frieder [4 ]
机构
[1] Univ Surrey, Ctr Vis Speech & Signal Proc CVSSP, Guildford GU2 7XH, Surrey, England
[2] UK Dementia Res Inst UK DRI, Care Res & Technol Ctr, London WC1E 6BT, England
[3] Imperial Coll London, Dept Brain Sci, London SW7 2BU, England
[4] Adobe Syst Engn GmbH, D-22767 Hamburg, Germany
关键词
Task analysis; Adaptation models; Training; Bayes methods; Modeling; Uncertainty; Gaussian distribution; Bayesian neural networks (BNNs); catastrophic forgetting; continual learning; incremental learning; uncertainty; SYSTEMS;
D O I
10.1109/TNNLS.2020.3017292
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Continual learning models allow them to learn and adapt to new changes and tasks over time. However, in continual and sequential learning scenarios, in which the models are trained using different data with various distributions, neural networks (NNs) tend to forget the previously learned knowledge. This phenomenon is often referred to as catastrophic forgetting. The catastrophic forgetting is an inevitable problem in continual learning models for dynamic environments. To address this issue, we propose a method, called continual Bayesian learning networks (CBLNs), which enables the networks to allocate additional resources to adapt to new tasks without forgetting the previously learned tasks. Using a Bayesian NN, CBLN maintains a mixture of Gaussian posterior distributions that are associated with different tasks. The proposed method tries to optimize the number of resources that are needed to learn each task and avoids an exponential increase in the number of resources that are involved in learning multiple tasks. The proposed method does not need to access the past training data and can choose suitable weights to classify the data points during the test time automatically based on an uncertainty criterion. We have evaluated the method on the MNIST and UCR time-series data sets. The evaluation results show that the method can address the catastrophic forgetting problem at a promising rate compared to the state-of-the-art models.
引用
收藏
页码:4243 / 4252
页数:10
相关论文
共 50 条
  • [21] Bayesian learning for recurrent neural networks
    Crucianu, M
    Boné, R
    de Beauville, JPA
    NEUROCOMPUTING, 2001, 36 (01) : 235 - 242
  • [22] OvA-INN: Continual Learning with Invertible Neural Networks
    Hocquet, Guillaume
    Bichler, Olivier
    Querlioz, Damien
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [23] Self-Organizing Incremental Neural Networks for Continual Learning
    Wiwatcharakoses, Chayut
    Berrar, Daniel
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 6476 - 6477
  • [24] Hybrid neural networks for continual learning inspired by corticohippocampal circuits
    Shi, Qianqian
    Liu, Faqiang
    Li, Hongyi
    Li, Guangyu
    Shi, Luping
    Zhao, Rong
    NATURE COMMUNICATIONS, 2025, 16 (01)
  • [25] Adaptive Neural Networks for Online Domain Incremental Continual Learning
    Gunasekara, Nuwan
    Gomes, Heitor
    Bifet, Albert
    Pfahringer, Bernhard
    DISCOVERY SCIENCE (DS 2022), 2022, 13601 : 89 - 103
  • [26] Continual Learning With Siamese Neural Networks for Sustainable Network Management
    El Yaacoub, Khalid
    Stenhammar, Oscar
    Ickin, Selim
    Vandikas, Konstantinos
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2024, 21 (03): : 2664 - 2674
  • [27] Continual learning of context-dependent processing in neural networks
    Guanxiong Zeng
    Yang Chen
    Bo Cui
    Shan Yu
    Nature Machine Intelligence, 2019, 1 : 364 - 372
  • [28] Continual Learning in Convolutional Neural Networks with Tensor Rank Updates
    Krol, Matt
    Hyder, Rakib
    Peechatt, Michael
    Prater-Bennette, Ashley
    Asif, M. Salman
    Markopoulos, Panos P.
    2024 IEEE 13RD SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING WORKSHOP, SAM 2024, 2024,
  • [29] Continual learning of context-dependent processing in neural networks
    Zeng, Guanxiong
    Chen, Yang
    Cui, Bo
    Yu, Shan
    NATURE MACHINE INTELLIGENCE, 2019, 1 (08) : 364 - 372
  • [30] Continual Learning using a Bayesian Nonparametric Dictionary of Weight Factors
    Mehta, Nikhil
    Liang, Kevin J.
    Verma, Vinay K.
    Carin, Lawrence
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130 : 100 - +