Find it if You Can: End-to-End Adversarial Erasing for Weakly-Supervised Semantic Segmentation

被引:3
作者
Stammes, Erik [1 ,2 ]
Runia, Tom F. H. [1 ]
Hofmann, Michael [2 ]
Ghafoorian, Mohsen [2 ]
机构
[1] Univ Amsterdam, Amsterdam, Netherlands
[2] TomTom, Amsterdam, Netherlands
来源
THIRTEENTH INTERNATIONAL CONFERENCE ON DIGITAL IMAGE PROCESSING (ICDIP 2021) | 2021年 / 11878卷
关键词
Semantic segmentation; adversarial learning; weak supervision;
D O I
10.1117/12.2599432
中图分类号
O43 [光学];
学科分类号
070207 ; 0803 ;
摘要
Semantic segmentation is a task that traditionally requires a large dataset of pixel-level ground truth labels, which is time-consuming and expensive to obtain. Recent advancements in the weakly-supervised setting show that reasonable performance can be obtained by using only image-level labels. Classification is often used as a proxy task to train a deep neural network from which attention maps are extracted. However, the classification task needs only the minimum evidence to make predictions, hence it focuses on the most discriminative object regions. To overcome this problem, we propose a novel formulation of adversarial erasing of the attention maps. In contrast to previous adversarial erasing methods, we optimize two networks with opposing loss functions, which eliminates the requirement of certain suboptimal strategies; for instance, having multiple training steps that complicate the training process or a weight sharing policy between networks operating on different distributions that might be suboptimal for performance. The proposed solution does not require saliency masks, instead it uses a regularization loss to prevent the attention maps from spreading to less discriminative object regions. Our experiments on the Pascal VOC dataset demonstrate that our adversarial approach increases segmentation performance by 2.1 mIoU compared to our baseline and by 1.0 mIoU compared to previous adversarial erasing approaches.
引用
收藏
页数:10
相关论文
共 44 条
[1]   Learning Pixel-level Semantic Affinity with Image-level Supervision forWeakly Supervised Semantic Segmentation [J].
Ahn, Jiwoon ;
Kwak, Suha .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :4981-4990
[2]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[3]   What's the Point: Semantic Segmentation with Point Supervision [J].
Bearman, Amy ;
Russakovsky, Olga ;
Ferrari, Vittorio ;
Fei-Fei, Li .
COMPUTER VISION - ECCV 2016, PT VII, 2016, 9911 :549-565
[4]   A Further Study on the Simulation of Cloud-Radiative Feedbacks in the ENSO Cycle in the Tropical Pacific with a Focus on the Asymmetry [J].
Chen, Lin ;
Sun, De-Zheng ;
Wang, Lu ;
Li, Tim .
ASIA-PACIFIC JOURNAL OF ATMOSPHERIC SCIENCES, 2019, 55 (03) :303-316
[5]   Attention-based Dropout Layer for Weakly Supervised Object Localization [J].
Choe, Junsuk ;
Shim, Hyunjung .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :2214-2223
[6]   BoxSup: Exploiting Bounding Boxes to Supervise Convolutional Networks for Semantic Segmentation [J].
Dai, Jifeng ;
He, Kaiming ;
Sun, Jian .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1635-1643
[7]   The Pascal Visual Object Classes (VOC) Challenge [J].
Everingham, Mark ;
Van Gool, Luc ;
Williams, Christopher K. I. ;
Winn, John ;
Zisserman, Andrew .
INTERNATIONAL JOURNAL OF COMPUTER VISION, 2010, 88 (02) :303-338
[8]  
Fan JS, 2020, AAAI CONF ARTIF INTE, V34, P10762
[9]   Associating Inter-image Salient Instances for Weakly Supervised Semantic Segmentation [J].
Fan, Ruochen ;
Hou, Qibin ;
Cheng, Ming-Ming ;
Yu, Gang ;
Martin, Ralph R. ;
Hu, Shi-Min .
COMPUTER VISION - ECCV 2018, PT IX, 2018, 11213 :371-388
[10]   Vision meets robotics: The KITTI dataset [J].
Geiger, A. ;
Lenz, P. ;
Stiller, C. ;
Urtasun, R. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2013, 32 (11) :1231-1237