Continual Learning for Object Classification: A Modular Approach

被引:1
作者
Turner, Daniel
Cardoso, Pedro J. S.
Rodrigues, Joao M. F. [1 ]
机构
[1] Univ Algarve, LARSyS, Faro, Portugal
来源
UNIVERSAL ACCESS IN HUMAN-COMPUTER INTERACTION. ACCESS TO MEDIA, LEARNING AND ASSISTIVE ENVIRONMENTS, UAHCI 2021, PT II | 2021年 / 12769卷
关键词
Neural Networks; Continual Learning; Catastrophic Forgetting;
D O I
10.1007/978-3-030-78095-1_39
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
A Human can immediately add new items to its set of known objects, whereas a computer, using traditional computer vision algorithms, would typically have to go almost back to the start and re-learn the all collection of objects (classes) from scratch. The reason the network must be re-trained is due to a phenomenon named Catastrophic Forgetting, where the changes made to the system during the acquisition of new knowledge brings about the loss of previous knowledge. In this paper, we explore the Continual Learning problem by proposing a way to deal with Catastrophic Forgetting. Our proposal includes a framework capable of learning new information without having to start from scratch and even "improve" its knowledge on what it already knows. With the above in mind, we present the Modular Dynamic Neural Network (MDNN), a network primarily made up of modular sub-networks that progressively grows in a tree shape and re-arranges itself as it learns continuously. The network is divided into two main blocks: (a) the feature extraction block, which is based on a ResNet50; and (b) the modular dynamic classification block, which is made up of sub-networks structured in such a way that its internal components function independently from one another. This structure allows that when new information is learned only specific sub-networks are altered in a way that old information is not forgotten. Tests show promising results with a set of ImageNet classes and also with a set of our own classes.
引用
收藏
页码:531 / 547
页数:17
相关论文
共 22 条
[1]  
Aljundi R, 2019, Arxiv, DOI arXiv:1806.05421
[2]   Broad Learning System: An Effective and Efficient Incremental Learning System Without the Need for Deep Architecture [J].
Chen, C. L. Philip ;
Liu, Zhulin .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (01) :10-24
[3]  
De Lange M, 2021, Arxiv, DOI arXiv:1909.08383
[4]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[5]   Adversarial Continual Learning [J].
Ebrahimi, Sayna ;
Meier, Franziska ;
Calandra, Roberto ;
Darrell, Trevor ;
Rohrbach, Marcus .
COMPUTER VISION - ECCV 2020, PT XI, 2020, 12356 :386-402
[6]   Recognizing Global Reservoirs From Landsat 8 Images: A Deep Learning Approach [J].
Fang, Weizhen ;
Wang, Cunguang ;
Chen, Xi ;
Wan, Wei ;
Li, Huan ;
Zhu, Siyu ;
Fang, Yu ;
Liu, Baojian ;
Hong, Yang .
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2019, 12 (09) :3168-3177
[7]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[8]  
Lomonaco V, 2017, PR MACH LEARN RES, V78
[9]  
van de Ven GM, 2019, Arxiv, DOI [arXiv:1904.07734, DOI 10.48550/ARXIV.1904.07734]
[10]   PackNet: Adding Multiple Tasks to a Single Network by Iterative Pruning [J].
Mallya, Arun ;
Lazebnik, Svetlana .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :7765-7773