FGNet: Fixation guidance network for salient object detection

被引:4
作者
Yuan, Junbin [1 ,2 ]
Xiao, Lifang [1 ]
Wattanachote, Kanoksak [2 ,3 ]
Xu, Qingzhen [1 ]
Luo, Xiaonan [4 ]
Gong, Yongyi [2 ,3 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou 510631, Peoples R China
[2] Guangdong Univ Foreign Studies, Intelligent Hlth & Visual Comp Lab, Guangzhou 510006, Peoples R China
[3] Guangdong Univ Foreign Studies, Sch Informat Sci & Technol, Guangzhou 510006, Peoples R China
[4] Guilin Univ Elect Technol, Sch Comp Sci & Informat Secur, Guilin 541004, Peoples R China
基金
美国国家科学基金会;
关键词
Salient object detection; Multi-task detection; Fixation guidance; Features interaction; CONVOLUTIONAL NEURAL-NETWORK; VISUAL-ATTENTION; MODEL;
D O I
10.1007/s00521-023-09028-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In challenging scenarios (e.g., small objects and cluttered backgrounds), most existing algorithms suffer from inconsistent results with human visual attention. Since fixation prediction can better model the human visual attention mechanism and has a strong correlation with salient objects. Inspired by this, we proposed a fixation guidance network (FGNet) for salient object detection, which innovatively used fixation prediction to guide both salient object detection and edge detection. Firstly, a multi-branch network structure was designed to achieve multi-task detection. Each branch unit significantly learned the extracted features to accomplish the correct prediction. Secondly, given the strong correlation between the fixation and salient objects, a fixation guidance module was employed to guide salient object detection and edge detection for obtaining more accurate detection results. Finally, to full use the complementary relationship between salient features and edge features, we proposed a multi-resolution feature interaction module to achieve mutual optimization within the same feature and between the different features for suppressing noise and enhancing their representations. The experimental results show that our proposed method performed better in challenging scenes and outperformed existing state-of-the-art algorithms in several metrics on four public benchmark datasets.
引用
收藏
页码:569 / 584
页数:16
相关论文
共 70 条
[1]  
Achanta R, 2008, LECT NOTES COMPUT SC, V5008, P66
[2]   Salient Object Detection: A Benchmark [J].
Borji, Ali ;
Cheng, Ming-Ming ;
Jiang, Huaizu ;
Li, Jia .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (12) :5706-5722
[3]  
Borji Ali, 2019, [Computational Visual Media, 计算可视媒体], V5, P117
[4]   State-of-the-Art in Visual Attention Modeling [J].
Borji, Ali ;
Itti, Laurent .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (01) :185-207
[5]  
Borji A, 2012, PROC CVPR IEEE, P478, DOI 10.1109/CVPR.2012.6247711
[6]  
Bruce NDB, 2005, INT C NEUR INF PROC, V18
[7]   Reverse Attention for Salient Object Detection [J].
Chen, Shuhan ;
Tan, Xiuli ;
Wang, Ben ;
Hu, Xuelong .
COMPUTER VISION - ECCV 2018, PT IX, 2018, 11213 :236-252
[8]   Reverse Attention-Based Residual Network for Salient Object Detection [J].
Chen, Shuhan ;
Tan, Xiuli ;
Wang, Ben ;
Lu, Huchuan ;
Hu, Xuelong ;
Fu, Yun .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 :3763-3776
[9]   Look, Perceive and Segment: Finding the Salient Objects in Images via Two-stream Fixation-Semantic CNNs [J].
Chen, Xiaowu ;
Zheng, Anlin ;
Li, Jia ;
Lu, Feng .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :1050-1058
[10]  
Chen ZY, 2020, AAAI CONF ARTIF INTE, V34, P10599