TCGNet: Type-Correlation Guidance for Salient Object Detection

被引:26
作者
Liu, Yi [1 ,2 ]
Zhou, Ling [1 ,3 ]
Wu, Gengshen [4 ]
Xu, Shoukun [1 ,3 ]
Han, Jungong [5 ]
机构
[1] Changzhou Univ, Sch Comp Sci & Artificial Intelligence, Aliyun Sch Big Data, Sch Software, Changzhou 213164, Jiangsu, Peoples R China
[2] Changzhou Univ, CNPC CZU Innovat Alliance, Changzhou 213164, Jiangsu, Peoples R China
[3] Changzhou Univ, Sch Software, Changzhou 213000, Jiangsu, Peoples R China
[4] City Univ Macau, Fac Data Sci, Macau 999078, Peoples R China
[5] Univ Sheffield, Dept Comp Sci, Sheffield S10 2TN, England
基金
中国国家自然科学基金;
关键词
Salient object detection; part-object relationship; capsule network;
D O I
10.1109/TITS.2023.3342811
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
Contrast and part-whole relations induced by deep neural networks like Convolutional Neural Networks (CNNs) and Capsule Networks (CapsNets) have been known as two types of semantic cues for deep salient object detection. However, few works pay attention to their complementary properties in the context of saliency prediction. In this paper, we probe into this issue and propose a Type-Correlation Guidance Network (TCGNet) for salient object detection. Specifically, a Multi-Type Cue Correlation (MTCC) covering CNNs and CapsNets is designed to extract the contrast and part-whole relational semantics, respectively. Using MTCC, two correlation matrices containing complementary information are computed with these two types of semantics. In return, these correlation matrices are used to guide the learning of the above semantics to generate better saliency cues. Besides, a Type Interaction Attention (TIA) is developed to interact semantics from CNNs and CapsNets for the aim of saliency prediction. Experiments and analysis on five benchmarks show the superiority of the proposed approach.
引用
收藏
页码:6633 / 6644
页数:12
相关论文
共 69 条
[1]  
Achanta R, 2009, PROC CVPR IEEE, P1597, DOI 10.1109/CVPRW.2009.5206596
[2]   Salient Object Detection: A Benchmark [J].
Borji, Ali ;
Sihite, Dicky N. ;
Itti, Laurent .
COMPUTER VISION - ECCV 2012, PT II, 2012, 7573 :414-429
[3]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[4]   A Highly Efficient Model to Study the Semantics of Salient Object Detection [J].
Cheng, Ming-Ming ;
Gao, Shang-Hua ;
Borji, Ali ;
Tan, Yong-Qiang ;
Lin, Zheng ;
Wang, Meng .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (11) :8006-8021
[5]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[6]   Where Does the Driver Look? Top-Down-Based Saliency Detection in a Traffic Driving Environment [J].
Deng, Tao ;
Yang, Kaifu ;
Li, Yongjie ;
Yan, Hongmei .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2016, 17 (07) :2051-2062
[7]   Egocentric Scene Understanding via Multimodal Spatial Rectifier [J].
Do, Tien ;
Vuong, Khiem ;
Park, Hyun Soo .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :2822-2831
[8]   Structure-measure: A New Way to Evaluate Foreground Maps [J].
Fan, Deng-Ping ;
Cheng, Ming-Ming ;
Liu, Yun ;
Li, Tao ;
Borji, Ali .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :4558-4567
[9]  
Fan DP, 2018, PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P698
[10]   Attentive Feedback Network for Boundary-Aware Salient Object Detection [J].
Feng, Mengyang ;
Lu, Huchuan ;
Ding, Errui .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1623-1632