Saliency and edge features-guided end-to-end network for salient object detection

被引:5
作者
Yang, Chen [1 ,2 ]
Xiao, Yang [1 ,2 ]
Chu, Lili [1 ]
Yu, Ziping [2 ,3 ]
Zhou, Jun [1 ]
Zheng, Huilong [2 ,3 ]
机构
[1] Liaoning Univ Technol, Sch Elect & Informat Engn, Jinzhou 121001, Peoples R China
[2] Chinese Acad Sci, Inst Engn Thermophys, Beijing 100190, Peoples R China
[3] Natl Key Lab Sci & Technol Adv Light Duty Gas Turb, Beijing 100190, Peoples R China
关键词
Vision transformer; Salient object detection; Edge detection; Multi-feature fusion; Set prediction; Random edge neighborhood sampling;
D O I
10.1016/j.eswa.2024.125016
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The rapid development of the Vision Transformer backbones has enabled the capture of feature information with global dependencies, leading to excellent performance in salient object detection tasks. However, it fails to adequately emphasize fine local features of edges, resulting in coarse and blurry edges in the final output. Therefore, this paper proposes a set prediction method for salient object detection. The approach enables the model to simultaneously consider both saliency (salient object) and edge features, achieving an end-to-end edge feature fusion. There is no need to design multiple complex branch structures and multiple training as with other methods. The model integrates random edge neighborhood sampling to enhance the recognition accuracy of local edge features in images. This approach addresses the weak perception of local features by Transformers and the issue encountered during actual training, where edge pixels typically occupy a much smaller proportion of the image compared to the background, resulting in insufficient learning of edge features by the model. The proposed end-to-end model in this paper fuses multiple features, including edge and salient objects. They are extracted within a unified framework while simultaneously outputting edge and salient object maps. Experimental results on six public datasets show that the proposed method significantly improves model performance benchmarks for detecting salient objects.
引用
收藏
页数:12
相关论文
共 56 条
  • [1] Carion Nicolas, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12346), P213, DOI 10.1007/978-3-030-58452-8_13
  • [2] Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation
    Chen, Liang-Chieh
    Zhu, Yukun
    Papandreou, George
    Schroff, Florian
    Adam, Hartwig
    [J]. COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 : 833 - 851
  • [3] Cheng B., 2021, arXiv, DOI [10.48550/arXiv.2107.06278, DOI 10.48550/ARXIV.2107.06278]
  • [4] HYPERSPECTRAL SALIENT OBJECT DETECTION USING EXTENDED MORPHOLOGY WITH CNN
    Chhapariya, Koushikey
    Buddhiraju, Krishna Mohan
    Kumar, Anil
    [J]. 2022 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2022), 2022, : 429 - 432
  • [5] BCNet: Bidirectional collaboration network for edge-guided salient object detection
    Dong, Bo
    Zhou, Yan
    Hu, Chuanfei
    Fu, Keren
    Chen, Geng
    [J]. NEUROCOMPUTING, 2021, 437 : 58 - 71
  • [6] Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
  • [7] UDNet: Uncertainty-aware deep network for salient object detection
    Fang, Yuming
    Zhang, Haiyan
    Yan, Jiebin
    Jiang, Wenhui
    Liu, Yang
    [J]. PATTERN RECOGNITION, 2023, 134
  • [8] Galteri L, 2018, INT C PATT RECOG, P3007
  • [9] Gu ZY, 2022, INT J COMPUT SCI ENG, V25, P504, DOI [10.1504/IJCSE.2022.10045026, 10.1504/IJCSE.2022.126253]
  • [10] Guolei Sun, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12347), P347, DOI 10.1007/978-3-030-58536-5_21