A Visual Representation-Guided Framework With Global Affinity for Weakly Supervised Salient Object Detection

被引:9
作者
Xu, Binwei [1 ]
Liang, Haoran [1 ]
Gong, Weihua [1 ]
Liang, Ronghua [1 ]
Chen, Peng [1 ]
机构
[1] Zhejiang Univ Technol, Coll Comp Sci & Technol, Hangzhou 310023, Peoples R China
关键词
Visualization; Task analysis; Image edge detection; Annotations; Semantics; Training; Object detection; General visual representation; global affinity; salient object detection; scribble; self-supervised transformer;
D O I
10.1109/TCSVT.2023.3284076
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Fully supervised salient object detection (SOD) methods have made considerable progress in performance, yet these models rely heavily on expensive pixel-wise labels. Recently, to achieve a trade-off between labeling burden and performance, scribble-based SOD methods have attracted increasing attention. Previous scribble-based models directly implement the SOD task only based on SOD training data with limited information, it is extremely difficult for them to understand the image and further achieve a superior SOD task. In this paper, we propose a simple yet effective framework guided by general visual representations with rich contextual semantic knowledge for scribble-based SOD. These general visual representations are generated by self-supervised learning based on large-scale unlabeled datasets. Our framework consists of a task-related encoder, a general visual module, and an information integration module to efficiently combine the general visual representations with task-related features to perform the SOD task based on understanding the contextual connections of images. Meanwhile, we propose a novel global semantic affinity loss to guide the model to perceive the global structure of the salient objects. Experimental results on five public benchmark datasets demonstrate that our method, which only utilizes scribble annotations without introducing any extra label, outperforms the state-of-theart weakly supervised SOD methods. Specifically, it outperforms the previous best scribble-based method on all datasets with an average gain of 5.5% for max f-measure, 5.8% for mean f-measure, 24% for MAE, and 3.1% for E-measure. Moreover, our method achieves comparable or even superior performance to the state-of-the-art fully supervised models.
引用
收藏
页码:248 / 259
页数:12
相关论文
共 50 条
  • [41] Global-Guided Weighted Enhancement for Salient Object Detection
    Yu, Jizhe
    Liu, Yu
    Wei, Hongkui
    Xu, Kaiping
    Li, Jiangquan
    Cao, Yifei
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT II, 2024, 15017 : 137 - 152
  • [42] Global Perception Network for Salient Object Detection in Remote Sensing Images
    Liu, Yu
    Zhang, Shanwen
    Wang, Zhen
    Zhao, Baoping
    Zou, Lincheng
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [43] Progressive Frame-Proposal Mining for Weakly Supervised Video Object Detection
    Han, Mingfei
    Wang, Yali
    Li, Mingjie
    Chang, Xiaojun
    Yang, Yi
    Qiao, Yu
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 1560 - 1573
  • [44] Negative Prototypes Guided Contrastive Learning for Weakly Supervised Object Detection
    Zhang, Yu
    Zhu, Chuang
    Yang, Guoqing
    Chen, Siqi
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT II, 2023, 14170 : 36 - 51
  • [45] High-Quality Proposals for Weakly Supervised Object Detection
    Cheng, Gong
    Yang, Junyu
    Gao, Decheng
    Guo, Lei
    Han, Junwei
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 5794 - 5804
  • [46] Efficient Weakly-Supervised Object Detection With Pseudo Annotations
    Yuan, Qingsheng
    Sun, Gang
    Liang, Jianming
    Leng, Biao
    [J]. IEEE ACCESS, 2021, 9 : 104356 - 104366
  • [47] CGMDRNet: Cross-Guided Modality Difference Reduction Network for RGB-T Salient Object Detection
    Chen, Gang
    Shao, Feng
    Chai, Xiongli
    Chen, Hangwei
    Jiang, Qiuping
    Meng, Xiangchao
    Ho, Yo-Sung
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (09) : 6308 - 6323
  • [48] Global contextual guided residual attention network for salient object detection
    Wang, Jun
    Zhao, Zhengyun
    Yang, Shangqin
    Chai, Xiuli
    Zhang, Wanjun
    Zhang, Miaohui
    [J]. APPLIED INTELLIGENCE, 2022, 52 (06) : 6208 - 6226
  • [49] Global contextual guided residual attention network for salient object detection
    Jun Wang
    Zhengyun Zhao
    Shangqin Yang
    Xiuli Chai
    Wanjun Zhang
    Miaohui Zhang
    [J]. Applied Intelligence, 2022, 52 : 6208 - 6226
  • [50] Salient Object Detection Based on Progressively Supervised Learning for Remote Sensing Images
    Zhang, Libao
    Ma, Jie
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2021, 59 (11): : 9682 - 9696