C2DFNet: Criss-Cross Dynamic Filter Network for RGB-D Salient Object Detection

被引:50
作者
Zhang, Miao [1 ]
Yao, Shunyu [2 ]
Hu, Beiqi [2 ]
Piao, Yongri [3 ]
Ji, Wei [4 ]
机构
[1] Dalian Univ Technol, DUT RU Int Sch Informat Sci & Software Engn, Key Lab Ubiquitous Network & Serv Software Liaonin, Dalian 116024, Peoples R China
[2] Dalian Univ Technol, Sch Software Technol, Dalian 116024, Peoples R China
[3] Dalian Univ Technol, Sch Informat & Commun Engn, Dalian 116024, Peoples R China
[4] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB T6G 2R3, Canada
基金
中国国家自然科学基金;
关键词
Dynamic filter; fusion network; RGB-D salient object detection;
D O I
10.1109/TMM.2022.3187856
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The ability to deal with intra and inter-modality features has been critical to the development of RGB-D salient object detection. While many works have advanced in leaps and bounds in this field, most existing methods have not taken their way down into the inherent differences between the RGB and depth data due to widely adopted conventional convolution in which fixed parameter kernels are applied during inference. To promote intra and inter-modality interaction conditioned on various scenarios, as RGB and depth data are processed independently and later fused interactively, we develop a new insight and a better model. In this paper, we introduce a criss-cross dynamic filter network by decoupling dynamic convolution. First, we propose a Model-specific Dynamic Enhanced Module (MDEM) that dynamically enhances the intra-modality features with global context guidance. Second, we propose a Scene-aware Dynamic Fusion Module (SDFM) to realize dynamic feature selection between two modalities. As a result, our model achieves accurate predictions of salient objects. Extensive experiments demonstrate that our method achieves competitive performance over 28 state-of-the-art RGB-D methods on 7 public datasets.
引用
收藏
页码:5142 / 5154
页数:13
相关论文
共 50 条
[31]   A cross-modal adaptive gated fusion generative adversarial network for RGB-D salient object detection [J].
Liu, Zhengyi ;
Zhang, Wei ;
Zhao, Peng .
NEUROCOMPUTING, 2020, 387 :210-220
[32]   Absolute and Relative Depth-Induced Network for RGB-D Salient Object Detection [J].
Kong, Yuqiu ;
Wang, He ;
Kong, Lingwei ;
Liu, Yang ;
Yao, Cuili ;
Yin, Baocai .
SENSORS, 2023, 23 (07)
[33]   MLBSNet: Mutual Learning and Boosting Segmentation Network for RGB-D Salient Object Detection [J].
Xia, Chenxing ;
Wang, Jingjing ;
Ge, Bing .
ELECTRONICS, 2024, 13 (14)
[34]   Cross-modal and multi-level feature refinement network for RGB-D salient object detection [J].
Gao, Yue ;
Dai, Meng ;
Zhang, Qing .
VISUAL COMPUTER, 2023, 39 (09) :3979-3994
[35]   DGFNet: Depth-Guided Cross-Modality Fusion Network for RGB-D Salient Object Detection [J].
Xiao, Fen ;
Pu, Zhengdong ;
Chen, Jiaqi ;
Gao, Xieping .
IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 :2648-2658
[36]   Cross-modal and multi-level feature refinement network for RGB-D salient object detection [J].
Yue Gao ;
Meng Dai ;
Qing Zhang .
The Visual Computer, 2023, 39 :3979-3994
[37]   CMA-SOD: cross-modal attention fusion network for RGB-D salient object detection [J].
Wang, Kexuan ;
Liu, Chenhua ;
Zhang, Rongfu .
VISUAL COMPUTER, 2024, :5135-5151
[38]   Transformer-based difference fusion network for RGB-D salient object detection [J].
Cui, Zhi-Qiang ;
Wang, Feng ;
Feng, Zheng-Yong .
JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
[39]   Swin Transformer-Based Edge Guidance Network for RGB-D Salient Object Detection [J].
Wang, Shuaihui ;
Jiang, Fengyi ;
Xu, Boqian .
SENSORS, 2023, 23 (21)
[40]   Guided residual network for RGB-D salient object detection with efficient depth feature learning [J].
Wang, Jian ;
Chen, Shuhan ;
Lv, Xiao ;
Xu, Xiuqi ;
Hu, Xuelong .
VISUAL COMPUTER, 2022, 38 (05) :1803-1814