Trigonometric feature learning for RGBD and RGBT image salient object detection

被引:0
作者
Huang, Liming [1 ,2 ]
Gong, Aojun [1 ]
机构
[1] Huazhong Univ Sci & Technol, Wuhan Natl Lab Optoelect, Wuhan, Peoples R China
[2] Univ Exeter, Fac Environm Sci & Econ, Dept Comp Sci, Exeter, England
关键词
Salient object detection; RGBD images; RGBT images; Feature mapping; Graph model; NETWORK;
D O I
10.1016/j.knosys.2024.112935
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
RGB-depth (RGBD) and RGB-thermal infrared (RGBT) images are integral to multi-modal salient object detection (SOD). Despite the impressive results of existing RGBD and RGBT SOD networks, two key areas remain in need of improvement. First, current methods rely heavily on feature fusion techniques, leading to overly complex and inflexible network architectures. Second, these methods lack a unified, adaptive strategy for combining features across different modalities or layers. To address these limitations, we propose a novel Trigonometric Feature Learning (TFL) strategy for generalized feature fusion in multi-modal SOD. Drawing inspiration from the trigonometric principles underlying vector operations, where the dot product of two vectors is the product of their magnitudes and the cosine of the angle between them, our TFL strategy maps features into graph space to compute the "cosine"mapping value for feature fusion. This cosine value dynamically adjusts based on feature attributes, enabling adaptive and effective fusion. To validate the TFL strategy, we design two network structures that use the TFL as the sole feature fusion mechanism for multi- modal SOD. Comparative evaluations against state-of-the-art methods demonstrate the strong performance of our networks, highlighting the unified and adaptive capabilities of the TFL strategy. The source code is available at https://github.com/huanglm-me/TFL-Net.git.
引用
收藏
页数:15
相关论文
共 83 条
  • [1] Achanta R, 2009, PROC CVPR IEEE, P1597, DOI 10.1109/CVPRW.2009.5206596
  • [2] 3-D Convolutional Neural Networks for RGB-D Salient Object Detection and Beyond
    Chen, Qian
    Zhang, Zhenxi
    Lu, Yanye
    Fu, Keren
    Zhao, Qijun
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 4309 - 4323
  • [3] Chen Q, 2021, AAAI CONF ARTIF INTE, V35, P1063
  • [4] Adaptive fusion network for RGB-D salient object detection
    Chen, Tianyou
    Xiao, Jin
    Hu, Xiaoguang
    Zhang, Guofeng
    Wang, Shaojie
    [J]. NEUROCOMPUTING, 2023, 522 : 152 - 164
  • [5] Depth-Induced Gap-Reducing Network for RGB-D Salient Object Detection: An Interaction, Guidance and Refinement Approach
    Cheng, Xiaolong
    Zheng, Xuan
    Pei, Jialun
    Tang, He
    Lyu, Zehua
    Chen, Chuanbo
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4253 - 4266
  • [6] Point-aware Interaction and CNN-induced Refinement Network for RGB-D Salient Object Detection
    Cong, Runmin
    Liu, Hongyu
    Zhang, Chen
    Zhang, Wei
    Zheng, Feng
    Song, Ran
    Kwong, Sam
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 406 - 416
  • [7] Does Thermal Really Always Matter for RGB-T Salient Object Detection?
    Cong, Runmin
    Zhang, Kepu
    Zhang, Chen
    Zheng, Feng
    Zhao, Yao
    Huang, Qingming
    Kwong, Sam
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 6971 - 6982
  • [8] Deng-Ping Fan, 2020, Computer Vision - ECCV 2020. 16th European Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12357), P275, DOI 10.1007/978-3-030-58610-2_17
  • [9] Fan DP, 2018, Arxiv, DOI arXiv:1805.10421
  • [10] Structure-measure: A New Way to Evaluate Foreground Maps
    Fan, Deng-Ping
    Cheng, Ming-Ming
    Liu, Yun
    Li, Tao
    Borji, Ali
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 4558 - 4567