Overcoming Catastrophic Forgetting in Continual Learning by Exploring Eigenvalues of Hessian Matrix

被引:13
|
作者
Kong, Yajing [1 ,2 ]
Liu, Liu [1 ,2 ]
Chen, Huanhuan [3 ]
Kacprzyk, Janusz [4 ]
Tao, Dacheng [1 ,2 ]
机构
[1] Univ Sydney, Sydney AI Ctr, Fac Engn, Darlington, NSW 2008, Australia
[2] Univ Sydney, Sch Comp Sci, Fac Engn, Darlington, NSW 2008, Australia
[3] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei 230027, Peoples R China
[4] Polish Acad Sci, Syst Res Inst, PL-01447 Warsaw, Poland
关键词
Task analysis; Convergence; Eigenvalues and eigenfunctions; Data models; Training; Upper bound; Loss measurement; Catastrophic forgetting; continual learning (CL); incremental learning; lifelong learning; NEURAL-NETWORKS; MEMORY;
D O I
10.1109/TNNLS.2023.3292359
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks tend to suffer performance deterioration on previous tasks when they are applied to multiple tasks sequentially without access to previous data. The problem is commonly known as catastrophic forgetting, a significant challenge in continual learning (CL). To overcome the catastrophic forgetting, regularization-based CL methods construct a regularization-based term, which can be considered as the approximation loss function of previous tasks, to penalize the update of parameters. However, the rigorous theoretical analysis of regularization-based methods is limited. Therefore, we theoretically analyze the forgetting and the convergence properties of regularization-based methods. The theoretical results demonstrate that the upper bound of the forgetting has a relationship with the maximum eigenvalue of the Hessian matrix. Hence, to decrease the upper bound of the forgetting, we propose eiGenvalues ExplorAtion Regularization-based (GEAR) method, which explores the geometric properties of the approximation loss of prior tasks regarding the maximum eigenvalue. Extensive experimental results demonstrate that our method mitigates catastrophic forgetting and outperforms existing regularization-based methods.
引用
收藏
页码:16196 / 16210
页数:15
相关论文
共 50 条
  • [41] Exploring Continual Learning of Compositional Generalization in NLI
    Fu, Xiyan
    Frank, Anette
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 912 - 932
  • [42] Exploring System Performance of Continual Learning for Mobile and Embedded Sensing Applications
    Kwon, Young D.
    Chauhan, Jagmohan
    Kumar, Abhishek
    Hui, Pan
    Mascolo, Cecilia
    2021 ACM/IEEE 6TH SYMPOSIUM ON EDGE COMPUTING (SEC 2021), 2021, : 319 - 332
  • [43] Hebbian learning rule restraining catastrophic forgetting in pulse neural network
    Motoki, M
    Hamagami, T
    Koakutsu, S
    Hirata, H
    ELECTRICAL ENGINEERING IN JAPAN, 2005, 151 (03) : 50 - 60
  • [44] Pseudo-rehearsal: Achieving deep reinforcement learning without catastrophic forgetting
    Atkinson, Craig
    McCane, Brendan
    Szymanski, Lech
    Robins, Anthony
    NEUROCOMPUTING, 2021, 428 : 291 - 307
  • [45] Investigating Catastrophic Forgetting of Deep Learning Models Within Office 31 Dataset
    Hidayaturrahman
    Trisetyarso, Agung
    Kartowisastro, Iman Herwidiana
    Budiharto, Widodo
    IEEE ACCESS, 2024, 12 : 138501 - 138509
  • [46] Bidirectional domain transfer knowledge distillation for catastrophic forgetting in federated learning with heterogeneous data
    Min, Qi
    Luo, Fei
    Dong, Wenbo
    Gu, Chunhua
    Ding, Weichao
    KNOWLEDGE-BASED SYSTEMS, 2025, 311
  • [47] Cuepervision: self-supervised learning for continuous domain adaptation without catastrophic forgetting
    Schutera, Mark
    Hafner, Frank M.
    Abhau, Jochen
    Hagenmeyer, Veit
    Mikut, Ralf
    Reischl, Markus
    IMAGE AND VISION COMPUTING, 2021, 106
  • [48] Handling catastrophic forgetting using cross-domain order in incremental deep learning
    Kumar, Ashutosh
    Agarwal, Sonali
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (02)
  • [49] Quick continual kernel learning on bounded memory space based on balancing between adaptation and forgetting
    Koichiro Yamauchi
    Evolving Systems, 2023, 14 : 437 - 460
  • [50] One Person, One Model, One World: Learning Continual User Representation without Forgetting
    Yuan, Fajie
    Zhang, Guoxiao
    Karatzoglou, Alexandros
    Jose, Joemon
    Kong, Beibei
    Li, Yudong
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 696 - 705