Spatial Continuity and Nonequal Importance in Salient Object Detection With Image-Category Supervision

被引:0
|
作者
Wu, Zhihao [1 ]
Liu, Chengliang [1 ]
Wen, Jie [1 ]
Xu, Yong [1 ,2 ]
Yang, Jian [3 ]
Li, Xuelong [4 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Shenzhen 518055, Peoples R China
[2] Pengcheng Lab, Shenzhen 518055, Peoples R China
[3] Nanjing Univ Sci & Technol, Dept Comp Sci & Engn, Nanjing 210094, Peoples R China
[4] Northwestern Polytech Univ, Sch Artificial Intelligence OPt & Elect iOPEN, Xian 710072, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Noise; Detectors; Transformers; Robustness; Object detection; Training; Feature extraction; Benchmark; robustness; salient object detection; weak supervision;
D O I
10.1109/TNNLS.2024.3436519
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the inefficiency of pixel-level annotations, weakly supervised salient object detection with image-category labels (WSSOD) has been receiving increasing attention. Previous works usually endeavor to generate high-quality pseudolabels to train the detectors in a fully supervised manner. However, we find that the detection performance is often limited by two types of noise contained in pseudolabels: 1) holes inside the object or at the edge and outliers in the background and 2) missing object portions and redundant surrounding regions. To mitigate the adverse effects caused by them, we propose local pixel correction (LPC) and key pixel attention (KPA), respectively, based on two key properties of desirable pseudolabels: 1) spatial continuity, meaning an object region consists of a cluster of adjacent points; and 2) nonequal importance, meaning pixels have different importance for training. Specifically, LPC fills holes and filters out outliers based on summary statistics of the neighborhood as well as its size. KPA directs the focus of training toward ambiguous pixels in multiple pseudolabels to discover more accurate saliency cues. To evaluate the effectiveness of our method, we design a simple yet strong baseline we call weakly supervised saliency detector with Transformer (WSSDT) and unify the proposed modules into WSSDT. Extensive experiments on five datasets demonstrate that our method significantly improves the baseline and outperforms all existing congeneric methods. Moreover, we establish the first benchmark to evaluate WSSOD robustness. The results show that our method can improve detection robustness as well. The code and robustness benchmark are available at https://github.com/Horatio9702/SCNI.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Salient object detection with image-level binary supervision
    Wang, Pengjie
    Liu, Yuxuan
    Cao, Ying
    Yang, Xin
    Luo, Yu
    Lu, Huchuan
    Liang, Zijian
    Lau, Rynson W. H.
    PATTERN RECOGNITION, 2022, 129
  • [2] Salient Object Detection With Importance Degree
    Umeki, Yo
    Funahashi, Isana
    Yoshida, Taichi
    Iwahashi, Masahiro
    IEEE ACCESS, 2020, 8 (08): : 147059 - 147069
  • [3] Weakly supervised salient object detection via image category annotation
    Zhang, Ruoqi
    Huang, Xiaoming
    Zhu, Qiang
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (12) : 21359 - 21381
  • [4] Pyramid Spatial Context Features for Salient Object Detection
    Li, Hui
    IEEE ACCESS, 2020, 8 : 88518 - 88526
  • [5] Approximately decoupled component supervision for salient object detection
    Yanhua Liang
    Guihe Qin
    Minghui Sun
    Jie Yan
    Zhonghan Zhang
    Applied Intelligence, 2022, 52 : 16117 - 16137
  • [6] SAC-Net: Spatial Attenuation Context for Salient Object Detection
    Hu, Xiaowei
    Fu, Chi-Wing
    Zhu, Lei
    Wang, Tianyu
    Heng, Pheng-Ann
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (03) : 1079 - 1090
  • [7] Approximately decoupled component supervision for salient object detection
    Liang, Yanhua
    Qin, Guihe
    Sun, Minghui
    Yan, Jie
    Zhang, Zhonghan
    APPLIED INTELLIGENCE, 2022, 52 (14) : 16117 - 16137
  • [8] View-Aware Salient Object Detection for 360° Omnidirectional Image
    Wu, Junjie
    Xia, Changqun
    Yu, Tianshu
    Li, Jia
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 6471 - 6484
  • [9] SALIENT OBJECT DETECTION IN IMAGE SEQUENCES VIA SPATIAL-TEMPORAL CUE
    Gan, Chuang
    Qin, Zengchang
    Xu, Jia
    Wan, Tao
    2013 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (IEEE VCIP 2013), 2013,
  • [10] Learning Complementary Spatial-Temporal Transformer for Video Salient Object Detection
    Liu, Nian
    Nan, Kepan
    Zhao, Wangbo
    Yao, Xiwen
    Han, Junwei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 10663 - 10673