Multi-scale feature correspondence and pseudo label retraining strategy for weakly supervised semantic segmentation

被引:0
作者
Wang, Weizheng [1 ]
Zhou, Lei [1 ]
Wang, Haonan [1 ]
机构
[1] Changsha Univ Sci & Technol, Sch Comp & Commun Engn, Changsha 410076, Peoples R China
基金
中国国家自然科学基金;
关键词
Weakly supervised semantic segmentation; Vision transformer; Multi-scale feature correspondence; Pseudo label retraining strategy;
D O I
10.1016/j.imavis.2024.105215
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, the performance of semantic segmentation using weakly supervised learning has significantly improved. Weakly supervised semantic segmentation (WSSS) that uses only image-level labels has received widespread attention, it employs Class Activation Maps (CAM) to generate pseudo labels. Compared to traditional use of pixel-level labels, this technique greatly reduces annotation costs by utilizing simpler and more readily available image-level annotations. Besides, due to the local perceptual ability of Convolutional Neural Networks (CNN), the generated CAM cannot activate the entire object area. Researchers have found that this CNN limitation can be compensated for by using Vision Transformer (ViT). However, ViT also introduces an over-smoothing problem. Recent research has made good progress in solving this issue, but when discussing CAM and its related segmentation predictions, it is easy to overlook their intrinsic information and the interrelationships between them. In this paper, we propose a Multi-Scale Feature Correspondence (MSFC) method. Our MSFC can obtain the feature correspondence of CAM and segmentation predictions at different scales, reextract useful semantic information from them, enhancing the network's learning of feature information and improving the quality of CAM. Moreover, to further improve the segmentation precision, we design a Pseudo Label Retraining Strategy (PLRS). This strategy refines the accuracy in local regions, elevates the quality of pseudo labels, and aims to enhance segmentation precision. Experimental results on the PASCAL VOC 2012 and MS COCO 2014 datasets show that our method achieves impressive performance among end-to-end WSSS methods.
引用
收藏
页数:11
相关论文
共 53 条
  • [31] CLIP is Also an Efficient Segmenter: A Text-Driven Approach for Weakly Supervised Semantic Segmentation
    Lin, Yuqi
    Chen, Minghao
    Wang, Wenxiao
    Wu, Boxi
    Li, Ke
    Lin, Binbin
    Liu, Haifeng
    He, Xiaofei
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 15305 - 15314
  • [32] Learning Self-supervised Low-Rank Network for Single-Stage Weakly and Semi-supervised Semantic Segmentation
    Pan, Junwen
    Zhu, Pengfei
    Zhang, Kaihua
    Cao, Bing
    Wang, Yu
    Zhang, Dingwen
    Han, Junwei
    Hu, Qinghua
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (05) : 1181 - 1195
  • [33] Max Pooling with Vision Transformers Reconciles Class and Shape in Weakly Supervised Semantic Segmentation
    Rossetti, Simone
    Zappia, Damiano
    Sanzari, Marta
    Schaerf, Marco
    Pirri, Fiora
    [J]. COMPUTER VISION - ECCV 2022, PT XXX, 2022, 13690 : 446 - 463
  • [34] Token Contrast for Weakly-Supervised Semantic Segmentation
    Ru, Lixiang
    Zheng, Hehang
    Zhan, Yibing
    Du, Bo
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 3093 - 3102
  • [35] Learning Affinity from Attention: End-to-End Weakly-Supervised Semantic Segmentation with Transformers
    Ru, Lixiang
    Zhan, Yibing
    Yu, Baosheng
    Du, Bo
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16825 - 16834
  • [36] ECS-Net: Improving Weakly Supervised Semantic Segmentation by Using Connections Between Class Activation Maps
    Sun, Kunyang
    Shi, Haoqing
    Zhang, Zhengming
    Huang, Yongming
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 7263 - 7272
  • [37] Vaswani A, 2017, ADV NEUR IN, V30
  • [38] Wang W., 2021, INT C LEARN REPR ICL
  • [39] Wang YD, 2020, PROC CVPR IEEE, P12272, DOI 10.1109/CVPR42600.2020.01229
  • [40] Wu F, 2024, P IEEECVF WINTER C A, P862