SEMEDA: Enhancing segmentation precision with semantic edge aware loss

被引:21
作者
Chen, Yifu [1 ]
Dapogny, Arnaud [1 ]
Cord, Matthieu [1 ]
机构
[1] Sorbonne Univ, UMR 7606, LIP6, 4 Pl Jussieu, Paris, France
关键词
Semantic segmentation; Loss function; Computer vision; NETWORKS;
D O I
10.1016/j.patcog.2020.107557
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Per-Pixel Cross entropy (PPCE) is a commonly used loss on semantic segmentation tasks. However, it suffers from a number of drawbacks. Firstly, PPCE only depends on the probability of the ground truth class since the latter is usually one-hot encoded. Secondly, PPCE treats all pixels independently and does not take the local structure into account. While perceptual losses (e.g. matching prediction and ground truth in the embedding space of a pre-trained VGG network) would theoretically address these concerns, does not constitute a practical solution as segmentation masks follow a distribution that differs largely from natural images. In this paper, we introduce a SEMantic EDge-Aware strategy (SEMEDA) to solve these issues. Inspired by perceptual losses, we propose to match the 'probability texture' of predicted segmentation mask and ground truth through a proxy network trained for semantic edge detection on the ground truth masks. Through thorough experimental validation on several datasets, we show that SEMEDA steadily improves the segmentation accuracy with negligible computational overhead and can be added with any popular segmentation networks in an end-to-end training framework. (c) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:13
相关论文
共 49 条
[1]  
[Anonymous], 2016, C NEUR INF PROC SYST
[2]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[3]   The Lovasz-Softmax loss: A tractable surrogate for the optimization of the intersection-over-union measure in neural networks [J].
Berman, Maxim ;
Triki, Amal Rannen ;
Blaschko, Matthew B. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4413-4421
[4]  
Chen L, 2015, P INT C LEARNING REP
[5]  
Chen LC., 2017, Computing Research Repository, V1706, P05587
[6]   Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Zhu, Yukun ;
Papandreou, George ;
Schroff, Florian ;
Adam, Hartwig .
COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 :833-851
[7]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[8]   Semantic Image Segmentation with Task-Specific Edge Detection Using CNNs and a Discriminatively Trained Domain Transform [J].
Chen, Liang-Chieh ;
Barron, Jonathan T. ;
Papandreou, George ;
Murphy, Kevin ;
Yuille, Alan L. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :4545-4554
[9]   Dual-force convolutional neural networks for accurate brain tumor segmentation [J].
Chen, Shengcong ;
Ding, Changxing ;
Liu, Minfeng .
PATTERN RECOGNITION, 2019, 88 :90-100
[10]   Deep Colorization [J].
Cheng, Zezhou ;
Yang, Qingxiong ;
Sheng, Bin .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :415-423