Bidirectional Self-Training with Multiple Anisotropic Prototypes for Domain Adaptive Semantic Segmentation

被引:13
|
作者
Lu, Yulei [1 ]
Luo, Yawei [1 ]
Zhang, Li [2 ]
Li, Zheyang [3 ]
Yang, Yi [1 ]
Xiao, Jun [1 ]
机构
[1] Zhejiang Univ, Hangzhou, Peoples R China
[2] Zhejiang Insigma Digital Technol Co Ltd, Hangzhou, Peoples R China
[3] Hikvis Res Inst, Hangzhou, Peoples R China
来源
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022 | 2022年
基金
中国国家自然科学基金; 浙江省自然科学基金;
关键词
Semantic Segmentation; Unsupervised Domain Adaptation; Gaussian Mixture Model; Self-training;
D O I
10.1145/3503161.3548225
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
A thriving trend for domain adaptive segmentation endeavors to generate the high-quality pseudo labels for target domain and retrain the segmentor on them. Under this self-training paradigm, some competitive methods have sought to the latent-space information, which establishes the feature centroids (a.k.a prototypes) of the semantic classes and determines the pseudo label candidates by their distances from these centroids. In this paper, we argue that the latent space contains more information to be exploited thus taking one step further to capitalize on it. Firstly, instead of merely using the source-domain prototypes to determine the target pseudo labels as most of the traditional methods do, we bidirectionally produce the target-domain prototypes to degrade those source features which might be too hard or disturbed for the adaptation. Secondly, existing attempts simply model each category as a single and isotropic prototype while ignoring the variance of the feature distribution, which could lead to the confusion of similar categories. To cope with this issue, we propose to represent each category with multiple and anisotropic prototypes via Gaussian Mixture Model, in order to fit the de facto distribution of source domain and estimate the likelihood of target samples based on the probability density. We apply our method on GTA5->Cityscapes and Synthia->Cityscapes tasks and achieve 61.2% and 62.8% respectively in terms of mean IoU, substantially outperforming other competitive self-training methods. Noticeably, in some categories which severely suffer from the categorical confusion such as "truck" and "bus", our method achieves 56.4% and 68.8% respectively, which further demonstrates the effectiveness of our design. The code and model are available at https://github.com/luyvlei/BiSMAPs.
引用
收藏
页码:1405 / 1415
页数:11
相关论文
共 50 条
  • [41] Adaptive Self-Training for Object Detection
    Vandeghen, Renaud
    Louppe, Gilles
    Van Droogenbroeck, Marc
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 914 - 923
  • [42] Improving Skin Lesion Segmentation with Self-Training
    Dzieniszewska, Aleksandra
    Garbat, Piotr
    Piramidowicz, Ryszard
    CANCERS, 2024, 16 (06)
  • [43] Single slice thigh CT muscle group segmentation with domain adaptation and self-training
    Yang, Qi
    Yu, Xin
    Lee, Ho Hin
    Cai, Leon Y.
    Xu, Kaiwen
    Bao, Shunxing
    Huo, Yuankai
    Moore, Ann Zenobia
    Makrogiannis, Sokratis
    Ferrucci, Luigi
    Landman, Bennett A.
    JOURNAL OF MEDICAL IMAGING, 2023, 10 (04)
  • [44] Domain Adaptation for Medical Image Segmentation Using Transformation-Invariant Self-training
    Ghamsarian, Negin
    Tejero, Javier Gamazo
    Marquez-Neila, Pablo
    Wolf, Sebastian
    Zinkernagel, Martin
    Schoeffmann, Klaus
    Sznitman, Raphael
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT I, 2023, 14220 : 331 - 341
  • [45] Unsupervised Domain Adaptation for Medical Image Segmentation via Self-Training of Early Features
    Sheikh, Rasha
    Schultz, Thomas
    INTERNATIONAL CONFERENCE ON MEDICAL IMAGING WITH DEEP LEARNING, VOL 172, 2022, 172 : 1096 - 1107
  • [46] Self-training adversarial learning for cross-domain retinal OCT fluid segmentation
    Li, Xiaohui
    Niu, Sijie
    Gao, Xizhan
    Zhou, Xueying
    Dong, Jiwen
    Zhao, Hui
    COMPUTERS IN BIOLOGY AND MEDICINE, 2023, 155
  • [47] Progressive Weighted Self-Training Ensemble for Multi-Type Skin Lesion Semantic Segmentation
    Lee, Cheolwon
    Yoo, Sangwook
    Kim, Semin
    Lee, Jongha
    IEEE ACCESS, 2022, 10 : 132376 - 132383
  • [48] A self-training end-to-end mask optimization framework based on semantic segmentation network*
    Xu, Hui
    Tang, Fuxin
    Qi, Pan
    Yuan, Ye
    Liang, Huaguo
    Huang, Zhengfeng
    INTEGRATION-THE VLSI JOURNAL, 2024, 96
  • [49] Towards Better Stability and Adaptability: Improve Online Self-Training for Model Adaptation in Semantic Segmentation
    Zhao, Dong
    Wang, Shuang
    Zang, Qi
    Quan, Dou
    Ye, Xiutiao
    Jiao, Licheng
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 11733 - 11743
  • [50] DAT: DOMAIN ADAPTIVE TRANSFORMER FOR DOMAIN ADAPTIVE SEMANTIC SEGMENTATION
    Park, Jinyoung
    Son, Minseok
    Lee, Sumin
    Kim, Changick
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 4183 - 4187