Multimodal feature adaptive fusion for anchor-free 3D object detection

被引:0
|
作者
Wu, Yanli [1 ]
Wang, Junyin [2 ]
Li, Hui [1 ]
Ai, Xiaoxue [1 ]
Li, Xiao [1 ]
机构
[1] Qingdao Univ Sci & Technol, Qingdao 266061, Peoples R China
[2] Wuhan Univ Technol, Wuhan 430070, Peoples R China
基金
国家重点研发计划;
关键词
Multimodal; 3D object detection; Point cloud; Anchor-free;
D O I
10.1007/s10489-025-06454-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
LiDAR and camera are two key sensors that provide mutually complementary information for 3D detection in autonomous driving. Existing multimodal detection methods often decorate the original point cloud data with camera features to complete the detection, ignoring the mutual fusion between camera features and point cloud features. In addition, ground points scanned by LiDAR in natural scenes usually interfere significantly with the detection results, and existing methods fail to address this problem effectively. We present a simple yet efficient anchor-free 3D object detection, which can better adapt to complex scenes through the adaptive fusion of multimodal features. First, we propose a fully convolutional bird's-eye view reconstruction module to sense ground map geometry changes, for improving the interference of ground points on detection results. Second, a multimodal feature adaptive fusion module with local awareness is designed to improve the mutual fusion of camera and point cloud features. Finally, we introduce a scale-aware mini feature pyramid networks (Mini-FPN) that can directly regress 3D bounding boxes from the augmented dense feature maps, boosting the network's ability to detect scale-varying objects, and we additionally construct a scene-adaptive single-stage 3D detector in an anchor-free manner. Extensive experiments on the KITTI and nuScenes datasets validate our method's competitive performance.
引用
收藏
页数:22
相关论文
共 50 条
  • [41] An improved anchor-free method for traffic scene object detection
    Ding, Tonghe
    Feng, Kaili
    Yan, Yejin
    Wei, Yanjun
    Li, Tianping
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (22) : 34703 - 34724
  • [42] MFFPN: an Anchor-Free Method for Patent Drawing Object Detection
    Chen, Yu-Hsien
    Chiu, Chih-Yi
    2023 18TH INTERNATIONAL CONFERENCE ON MACHINE VISION AND APPLICATIONS, MVA, 2023,
  • [43] Domain Adaptation of Anchor-Free object detection for urban traffic
    Yu, Xiaoyong
    Lu, Xiaoqiang
    NEUROCOMPUTING, 2024, 582
  • [44] An improved anchor-free method for traffic scene object detection
    Tonghe Ding
    Kaili Feng
    Yejin Yan
    Yanjun Wei
    Tianping Li
    Multimedia Tools and Applications, 2023, 82 : 34703 - 34724
  • [45] Anchor-free Proposal Generation Network for Efficient Object Detection
    Nguyen, Hoanh
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2023, 14 (04) : 327 - 335
  • [46] One-Stage Anchor-Free 3D Vehicle Detection from LiDAR Sensors
    Li, Hao
    Zhao, Sanyuan
    Zhao, Wenjun
    Zhang, Libin
    Shen, Jianbing
    SENSORS, 2021, 21 (08)
  • [47] 3D Anchor-Free Lesion Detector on Computed Tomography Scans
    Zhang, Ning
    Wang, Dechun
    Sun, Xinzi
    Zhang, Pengfei
    Zhang, Chenxi
    Cao, Yu
    Liu, Benyuan
    2019 FIRST INTERNATIONAL CONFERENCE ON TRANSDISCIPLINARY AI (TRANSAI 2019), 2019, : 48 - 51
  • [48] Multi-feature Fusion VoteNet for 3D Object Detection
    Wang, Zhoutao
    Xie, Qian
    Wei, Mingqiang
    Long, Kun
    Wang, Jun
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2022, 18 (01)
  • [49] Adaptive learning point cloud and image diversity feature fusion network for 3D object detection
    Yan, Weiqing
    Liu, Shile
    Liu, Hao
    Yue, Guanghui
    Wang, Xuan
    Song, Yongchao
    Xu, Jindong
    COMPLEX & INTELLIGENT SYSTEMS, 2024, 10 (02) : 2825 - 2837
  • [50] Adaptive learning point cloud and image diversity feature fusion network for 3D object detection
    Weiqing Yan
    Shile Liu
    Hao Liu
    Guanghui Yue
    Xuan Wang
    Yongchao Song
    Jindong Xu
    Complex & Intelligent Systems, 2024, 10 : 2825 - 2837