Continual Attentive Fusion for Incremental Learning in Semantic Segmentation

被引:16
作者
Yang, Guanglei [1 ]
Fini, Enrico [2 ]
Xu, Dan [3 ]
Rota, Paolo [2 ]
Ding, Mingli [1 ]
Tang, Hao [4 ]
Alameda-Pineda, Xavier [5 ]
Ricci, Elisa [6 ,7 ]
机构
[1] Harbin Inst Technol HIT, Sch Instrument Sci & Engn, Harbin 150001, Peoples R China
[2] Univ Trento, Dept Informat Engn & Comp Sci, I-38123 Povo, Italy
[3] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Hong Kong 999077, Peoples R China
[4] Swiss Fed Inst Technol, Dept Informat Technol & Elect Engn, CH-8092 Zurich, Switzerland
[5] INRIA, RobotLearn Grp, F-38330 Montbonnot St Martin, France
[6] Univ Trento, Dept Informat Engn & Comp Sci, I-38123 Povo, Italy
[7] Fdn Bruno Kessler, Deep Visual Learning Grp, I-38123 Trento, Italy
基金
欧盟地平线“2020”;
关键词
Task analysis; Semantics; Image segmentation; Tensors; Feature extraction; Deep learning; Training; Incremental learning; knowledge distillation; semantic segmentation;
D O I
10.1109/TMM.2022.3167555
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Over the past years, semantic segmentation, similar to many other tasks in computer vision, has benefited from the progress in deep neural networks, resulting in significantly improved performance. However, deep architectures trained with gradient-based techniques suffer from catastrophic forgetting, which is the tendency to forget previously learned knowledge while learning new tasks. Aiming at devising strategies to counteract this effect, incremental learning approaches have gained popularity over the past years. However, the first incremental learning methods for semantic segmentation appeared only recently. While effective, these approaches do not account for a crucial aspect in pixel-level dense prediction problems, i.e., the role of attention mechanisms. To fill this gap, in this paper, we introduce a novel attentive feature distillation approach to mitigate catastrophic forgetting while accounting for semantic spatial- and channellevel dependencies. Furthermore, we propose a continual attentive fusion structure, which takes advantage of the attention learned from the new and the old tasks while learning features for the new task. Finally, we also introduce a novel strategy to account for the background class in the distillation loss, thus preventing biased predictions. We demonstrate the effectiveness of our approach with an extensive evaluation on Pascal-VOC 2012 and ADE20 K, setting a new state of the art.
引用
收藏
页码:3841 / 3854
页数:14
相关论文
共 65 条
  • [1] Memory Aware Synapses: Learning What (not) to Forget
    Aljundi, Rahaf
    Babiloni, Francesca
    Elhoseiny, Mohamed
    Rohrbach, Marcus
    Tuytelaars, Tinne
    [J]. COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 : 144 - 161
  • [2] CASTRO FM, 2018, P EUR C COMPUT VIS, P320
  • [3] Modeling the Background for Incremental Learning in Semantic Segmentation
    Cermelli, Fabio
    Mancini, Massimiliano
    Bulo, Samuel Rota
    Ricci, Elisa
    Caputo, Barbara
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, : 9230 - 9239
  • [4] Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence
    Chaudhry, Arslan
    Dokania, Puneet K.
    Ajanthan, Thalaiyasingam
    Torr, Philip H. S.
    [J]. COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 : 556 - 572
  • [5] Chen LC., 2017, RETHINKING ATROUS CO
  • [6] CaMap: Camera-based Map Manipulation on Mobile Devices
    Chen, Liang
    Chen, Dongyi
    [J]. PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND APPLICATION ENGINEERING (CSAE2018), 2018,
  • [7] Attention to Scale: Scale-aware Semantic Image Segmentation
    Chen, Liang-Chieh
    Yang, Yi
    Wang, Jiang
    Xu, Wei
    Yuille, Alan L.
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 3640 - 3649
  • [8] Feature Estimations Based Correlation Distillation for Incremental Image Retrieval
    Chen, Wei
    Liu, Yu
    Pu, Nan
    Wang, Weiping
    Liu, Li
    Lew, Michael S.
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 1844 - 1856
  • [9] Chorowski J. K, 2015, ADV NEURAL INFORM PR, V1, P577, DOI DOI 10.1016/0167-739X(94)90007-8
  • [10] DELANGE M, 2019, CONTINUAL LEARNING A