Hierarchical and Interactive Refinement Network for Edge-Preserving Salient Object Detection

被引:28
作者
Zhou, Sanping [1 ]
Wang, Jinjun [1 ]
Wang, Le [1 ]
Zhang, Jimuyang [2 ]
Wang, Fei [3 ]
Huang, Dong [2 ]
Zheng, Nanning [1 ]
机构
[1] Xi An Jiao Tong Univ, Inst Artificial Intelligence & Robot, Xian 710049, Peoples R China
[2] Carnegie Mellon Univ, Inst Robot, Pittsburgh, PA 15213 USA
[3] Xi An Jiao Tong Univ, Sch Software Engn, Xian 710049, Peoples R China
基金
中国国家自然科学基金;
关键词
Image edge detection; Object detection; Feature extraction; Inference algorithms; Training; Prediction algorithms; Semantics; Salient object detection; edge-guided inference; Hierarchical and Interactive Refinement Network; NEURAL-NETWORK;
D O I
10.1109/TIP.2020.3027992
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Salient object detection has undergone a very rapid development with the blooming of Deep Neural Network (DNN), which is usually taken as an important preprocessing procedure in various computer vision tasks. However, the down-sampling operations, such as pooling and striding, always make the final predictions blurred at edges, which has seriously degenerated the performance of salient object detection. In this paper, we propose a simple yet effective approach, i.e., Hierarchical and Interactive Refinement Network (HIRN), to preserve the edge structures in detecting salient objects. In particular, a novel multi-stage and dual-path network structure is designed to estimate the salient edges and regions from the low-level and high-level feature maps, respectively. As a result, the predicted regions will become more accurate by enhancing the weak responses at edges, while the predicted edges will become more semantic by suppressing the false positives in background. Once the salient maps of edges and regions are obtained at the output layers, a novel edge-guided inference algorithm is introduced to further filter the resulting regions along the predicted edges. Extensive experiments on several benchmark datasets have been conducted, in which the results show that our method significantly outperforms a variety of state-of-the-art approaches.
引用
收藏
页码:1 / 14
页数:14
相关论文
共 72 条
  • [1] [Anonymous], IEEE T PATTERN ANAL, DOI DOI 10.1007/978-3-642-14267-3_2
  • [2] Reverse Attention for Salient Object Detection
    Chen, Shuhan
    Tan, Xiuli
    Wang, Ben
    Hu, Xuelong
    [J]. COMPUTER VISION - ECCV 2018, PT IX, 2018, 11213 : 236 - 252
  • [3] Global Contrast based Salient Region Detection
    Cheng, Ming-Ming
    Zhang, Guo-Xin
    Mitra, Niloy J.
    Huang, Xiaolei
    Hu, Shi-Min
    [J]. 2011 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2011, : 409 - 416
  • [4] Deng ZJ, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P684
  • [5] Attentive Feedback Network for Boundary-Aware Salient Object Detection
    Feng, Mengyang
    Lu, Huchuan
    Ding, Errui
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 1623 - 1632
  • [6] Gao WS, 2010, INT CONF COMP SCI, P67, DOI 10.1109/ICCSIT.2010.5563693
  • [7] Edge-Aware Convolution Neural Network Based Salient Object Detection
    Guan, Wenlong
    Wang, Tiantian
    Qi, Jinqing
    Zhang, Lihe
    Lu, Huchuan
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2019, 26 (01) : 114 - 118
  • [8] Co-teaching: Robust Training of Deep Neural Networks with Extremely Noisy Labels
    Han, Bo
    Yao, Quanming
    Yu, Xingrui
    Niu, Gang
    Xu, Miao
    Hu, Weihua
    Tsang, Ivor W.
    Sugiyama, Masashi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [9] SalNet: Edge Constraint Based End-to-End Model for Salient Object Detection
    Han, Le
    Li, Xuelong
    Dong, Yongsheng
    [J]. PATTERN RECOGNITION AND COMPUTER VISION (PRCV 2018), PT IV, 2018, 11259 : 186 - 198
  • [10] Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1026 - 1034