A Visual Representation-Guided Framework With Global Affinity for Weakly Supervised Salient Object Detection

被引:9
作者
Xu, Binwei [1 ]
Liang, Haoran [1 ]
Gong, Weihua [1 ]
Liang, Ronghua [1 ]
Chen, Peng [1 ]
机构
[1] Zhejiang Univ Technol, Coll Comp Sci & Technol, Hangzhou 310023, Peoples R China
关键词
Visualization; Task analysis; Image edge detection; Annotations; Semantics; Training; Object detection; General visual representation; global affinity; salient object detection; scribble; self-supervised transformer;
D O I
10.1109/TCSVT.2023.3284076
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Fully supervised salient object detection (SOD) methods have made considerable progress in performance, yet these models rely heavily on expensive pixel-wise labels. Recently, to achieve a trade-off between labeling burden and performance, scribble-based SOD methods have attracted increasing attention. Previous scribble-based models directly implement the SOD task only based on SOD training data with limited information, it is extremely difficult for them to understand the image and further achieve a superior SOD task. In this paper, we propose a simple yet effective framework guided by general visual representations with rich contextual semantic knowledge for scribble-based SOD. These general visual representations are generated by self-supervised learning based on large-scale unlabeled datasets. Our framework consists of a task-related encoder, a general visual module, and an information integration module to efficiently combine the general visual representations with task-related features to perform the SOD task based on understanding the contextual connections of images. Meanwhile, we propose a novel global semantic affinity loss to guide the model to perceive the global structure of the salient objects. Experimental results on five public benchmark datasets demonstrate that our method, which only utilizes scribble annotations without introducing any extra label, outperforms the state-of-theart weakly supervised SOD methods. Specifically, it outperforms the previous best scribble-based method on all datasets with an average gain of 5.5% for max f-measure, 5.8% for mean f-measure, 24% for MAE, and 3.1% for E-measure. Moreover, our method achieves comparable or even superior performance to the state-of-the-art fully supervised models.
引用
收藏
页码:248 / 259
页数:12
相关论文
共 50 条
  • [31] CoGANet: Co-Guided Attention Network for Salient Object Detection
    Zhao, Yufei
    Song, Yong
    Li, Guoqi
    Huang, Yi
    Bai, Yashuo
    Zhou, Ya
    Hao, Qun
    IEEE PHOTONICS JOURNAL, 2022, 14 (04):
  • [32] Self-progress aggregate learning for weakly supervised salient object detection
    Sun, Wanchun
    Feng, Xin
    Liu, Jingyao
    Ma, Hui
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2023, 34 (06)
  • [33] To Be Critical: Self-calibrated Weakly Supervised Learning for Salient Object Detection
    Wang, Jian
    Liu, Tingwei
    Zhang, Miao
    Piao, Yongri
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XI, 2024, 14435 : 184 - 198
  • [34] Noise-Sensitive Adversarial Learning for Weakly Supervised Salient Object Detection
    Piao, Yongri
    Wu, Wei
    Zhang, Miao
    Jiang, Yongyao
    Lu, Huchuan
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 2888 - 2897
  • [35] Contrastive Proposal Extension With LSTM Network for Weakly Supervised Object Detection
    Lv, Pei
    Hu, Suqi
    Hao, Tianran
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 6879 - 6892
  • [36] DNA: Deeply Supervised Nonlinear Aggregation for Salient Object Detection
    Liu, Yun
    Cheng, Ming-Ming
    Zhang, Xin-Yu
    Nie, Guang-Yu
    Wang, Meng
    IEEE TRANSACTIONS ON CYBERNETICS, 2022, 52 (07) : 6131 - 6142
  • [37] CGFNet: Cross-Guided Fusion Network for RGB-T Salient Object Detection
    Wang, Jie
    Song, Kechen
    Bao, Yanqi
    Huang, Liming
    Yan, Yunhui
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (05) : 2949 - 2961
  • [38] A Weakly-Supervised Cross-Domain Query Framework for Video Camouflage Object Detection
    Lu, Zelin
    Xie, Liang
    Zhao, Xing
    Xu, Binwei
    Liang, Haoran
    Liang, Ronghua
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (02) : 1506 - 1518
  • [39] Spatial Attention-Guided Light Field Salient Object Detection Network With Implicit Neural Representation
    Zheng, Xin
    Li, Zhengqu
    Liu, Deyang
    Zhou, Xiaofei
    Shan, Caifeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (12) : 12437 - 12449
  • [40] Self-Supervised Pretraining With Multimodality Representation Enhancement for Salient Object Detection in RGB-D Images
    Gao, Lina
    Liu, Bing
    Fu, Ping
    Xu, Mingzhu
    Zhang, Yonggang
    Huang, Yulong
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2025, 74