Modular Dynamic Neural Network: A Continual Learning Architecture

被引:4
作者
Turner, Daniel [1 ]
Cardoso, Pedro J. S. [1 ]
Rodrigues, Joao M. F. [1 ]
机构
[1] Univ Algarve, LARSYS & ISE, P-8005139 Faro, Portugal
来源
APPLIED SCIENCES-BASEL | 2021年 / 11卷 / 24期
关键词
continual learning; neural networks; catastrophic forgetting; object recognition;
D O I
10.3390/app112412078
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Learning to recognize a new object after having learned to recognize other objects may be a simple task for a human, but not for machines. The present go-to approaches for teaching a machine to recognize a set of objects are based on the use of deep neural networks (DNN). So, intuitively, the solution for teaching new objects on the fly to a machine should be DNN. The problem is that the trained DNN weights used to classify the initial set of objects are extremely fragile, meaning that any change to those weights can severely damage the capacity to perform the initial recognitions; this phenomenon is known as catastrophic forgetting (CF). This paper presents a new (DNN) continual learning (CL) architecture that can deal with CF, the modular dynamic neural network (MDNN). The presented architecture consists of two main components: (a) the ResNet50-based feature extraction component as the backbone; and (b) the modular dynamic classification component, which consists of multiple sub-networks and progressively builds itself up in a tree-like structure that rearranges itself as it learns over time in such a way that each sub-network can function independently. The main contribution of the paper is a new architecture that is strongly based on its modular dynamic training feature. This modular structure allows for new classes to be added while only altering specific sub-networks in such a way that previously known classes are not forgotten. Tests on the CORe50 dataset showed results above the state of the art for CL architectures.
引用
收藏
页数:21
相关论文
共 36 条
[1]  
Aljundi Rahaf, 2018, ARXIV180605421
[2]  
[Anonymous], 2009, Master's Thesis
[3]  
Chatfield K, 2014, ARXIV
[4]   Broad Learning System: An Effective and Efficient Incremental Learning System Without the Need for Deep Architecture [J].
Chen, C. L. Philip ;
Liu, Zhulin .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (01) :10-24
[5]  
Chen Z., 2018, SYNTH LECT ARTIF INT, V12, P1
[6]   A Continual Learning Survey: Defying Forgetting in Classification Tasks [J].
De Lange, Matthias ;
Aljundi, Rahaf ;
Masana, Marc ;
Parisot, Sarah ;
Jia, Xu ;
Leonardis, Ales ;
Slabaugh, Greg ;
Tuytelaars, Tinne .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (07) :3366-3385
[7]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[8]  
Ebrahimi S., 2020, P ECCV, V12356, P386
[9]  
Fang W., 2018, P AGU FALL M WASH DC
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778