Exploring High-quality Target Domain Information for Unsupervised Domain Adaptive Semantic Segmentation

被引:11
作者
Li, Junjie [1 ]
Wang, Zilei [1 ]
Gao, Yuan [1 ]
Hu, Xiaoming [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
来源
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022 | 2022年
基金
中国国家自然科学基金;
关键词
unsupervised domain adaptive semantic segmentation; contrastive learning; pseudo labels;
D O I
10.1145/3503161.3548114
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In unsupervised domain adaptive (UDA) semantic segmentation, the distillation based methods are currently dominant in performance. However, the distillation technique requires complicate multi-stage process and many training tricks. In this paper, we propose a simple yet effective method that can achieve competitive performance to the advanced distillation methods. Our core idea is to fully explore the target-domain information from the views of boundaries and features. First, we propose a novel mix-up strategy to generate high-quality target-domain boundaries with ground-truth labels. Different from the source-domain boundaries in previous works, we select the high-confidence target-domain areas and then paste them to the source-domain images. Such a strategy can generate the object boundaries in target domain (edge of target-domain object areas) with the correct labels. Consequently, the boundary information of target domain can be effectively captured by learning on the mixed-up samples. Second, we design a multi-level contrastive loss to improve the representation of target-domain data, including pixel-level and prototypelevel contrastive learning. By combining two proposed methods, more discriminative features can be extracted and hard object boundaries can be better addressed for the target domain. The experimental results on two commonly adopted benchmarks (i.e., GTA5. Cityscapes and SYNTHIA. Cityscapes) show that our method achieves competitive performance to complicated distillation methods. Notably, for the SYNTHIA. Cityscapes scenario, our method achieves the state-of-the-art performance with 57.8% mIoU and 64.6% mIoU on 16 classes and 13 classes. Code is available at https://github.com/ljjcoder/EHTDI.
引用
收藏
页码:5237 / 5245
页数:9
相关论文
共 45 条
[1]   Self-supervised Augmentation Consistency for Adapting Semantic Segmentation [J].
Araslanov, Nikita ;
Roth, Stefan .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :15379-15389
[2]  
Caron M, 2020, ADV NEUR IN, V33
[3]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[4]  
Chen MH, 2020, AAAI CONF ARTIF INTE, V34, P3521
[5]   Domain Adaptation for Semantic Segmentation with Maximum Squares Loss [J].
Chen, Minghao ;
Xue, Hongyang ;
Cai, Deng .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :2090-2099
[6]  
Chen T., 2020, Advances in neural information processing systems, P22243
[7]  
Chen Ting, 2019, PMLR
[8]  
Cheng Yiting, 2021, ICCV, P9082
[9]   The Cityscapes Dataset for Semantic Urban Scene Understanding [J].
Cordts, Marius ;
Omran, Mohamed ;
Ramos, Sebastian ;
Rehfeld, Timo ;
Enzweiler, Markus ;
Benenson, Rodrigo ;
Franke, Uwe ;
Roth, Stefan ;
Schiele, Bernt .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3213-3223
[10]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848