Continual prune-and-select: class-incremental learning with specialized subnetworks

被引:0
作者
Aleksandr Dekhovich
David M.J. Tax
Marcel H.F Sluiter
Miguel A. Bessa
机构
[1] Delft University of Technology,Department of Materials Science and Engineering
[2] Delft University of Technology,Pattern Recognition and Bioinformatics Laboratory
[3] Brown University,School of Engineering
来源
Applied Intelligence | 2023年 / 53卷
关键词
Continual learning; Class-incremental learning; Sparse network representation; Catastrophic forgetting;
D O I
暂无
中图分类号
学科分类号
摘要
The human brain is capable of learning tasks sequentially mostly without forgetting. However, deep neural networks (DNNs) suffer from catastrophic forgetting when learning one task after another. We address this challenge considering a class-incremental learning scenario where the DNN sees test data without knowing the task from which this data originates. During training, Continual Prune-and-Select (CP&S) finds a subnetwork within the DNN that is responsible for solving a given task. Then, during inference, CP&S selects the correct subnetwork to make predictions for that task. A new task is learned by training available neuronal connections of the DNN (previously untrained) to create a new subnetwork by pruning, which can include previously trained connections belonging to other subnetwork(s) because it does not update shared connections. This enables to eliminate catastrophic forgetting by creating specialized regions in the DNN that do not conflict with each other while still allowing knowledge transfer across them. The CP&S strategy is implemented with different subnetwork selection strategies, revealing superior performance to state-of-the-art continual learning methods tested on various datasets (CIFAR-100, CUB-200-2011, ImageNet-100 and ImageNet-1000). In particular, CP&S is capable of sequentially learning 10 tasks from ImageNet-1000 keeping an accuracy around 94% with negligible forgetting, a first-of-its-kind result in class-incremental learning. To the best of the authors’ knowledge, this represents an improvement in accuracy above 10% when compared to the best alternative method.
引用
收藏
页码:17849 / 17864
页数:15
相关论文
共 29 条
[1]  
French RM(1999)Catastrophic forgetting in connectionist networks Trends Cognit Sci 3 128-135
[2]  
Lerner Y(2011)Topographic mapping of a hierarchy of temporal receptive windows using a narrated story J Neurosci 31 2906-2915
[3]  
Honey CJ(2017)How we transmit memories to other brains: constructing shared neural representations via communication Cerebral cortex 27 4988-5000
[4]  
Silbert LJ(1990)Morphometric study of human cerebral cortex development Neuropsychologia 28 517-527
[5]  
Hasson U(2003)The cost of cortical computation Current Biol 13 493-497
[6]  
Zadbood A(2001)An energy budget for signaling in the grey matter of the brain J Cerebral Blood Flow Metabolism 21 1133-1145
[7]  
Chen J(2017)Learning without forgetting IEEE Trans Pattern Anal Mach Intell 40 2935-2947
[8]  
Leong YC(2017)Overcoming catastrophic forgetting in neural networks Proc National Acad Sci 114 3521-3526
[9]  
Norman KA(2021)Spacenet: make free space for continual learning Neurocomputing 439 1-11
[10]  
Hasson U(undefined)undefined undefined undefined undefined-undefined