Millimeter-Wave Radar and Camera Fusion for Multiscenario Object Detection on USVs

被引:2
作者
He, Xin [1 ,2 ]
Wu, Defeng [1 ,2 ]
Wu, Dongjie [1 ,2 ]
You, Zheng [1 ,2 ]
Zhong, Shangkun [1 ,2 ]
Liu, Qijun [1 ,2 ]
机构
[1] Jimei Univ, Fujian Inst Innovat Marine Equipment Detect & Remf, Sch Marine Engn, Xiamen 361021, Fujian, Peoples R China
[2] Jimei Univ, Fujian Prov Key Lab Naval Architecture & Ocean Eng, Xiamen 361021, Fujian, Peoples R China
基金
中国国家自然科学基金;
关键词
Radar; Cameras; Radar imaging; Radar detection; Sensors; Feature extraction; Millimeter wave radar; Deep learning; fusion mixture with AFPN (FMA)-fully convolutional one-stage (FCOS); multiscenario; object detection; sensor fusion; unmanned surface vehicle (USV); NAVIGATION;
D O I
10.1109/JSEN.2024.3444826
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Accurate object detection is fundamental for unmanned surface vehicles (USVs) to achieve intelligent perception. This article proposes an object detection network that integrates millimeter-wave radar and a camera. The method utilizes the complementary advantages of millimeter-wave radar and camera data modalities to realize multiscenario object detection for USVs applications. To address the drawback of sparse point clouds in millimeter-wave radar and improve the suboptimal performance of the camera in adverse weather conditions and small object detection, as well as to effectively utilize the features of both millimeter-wave radar and camera, a multisensor deep learning fusion object detection network [fusion mixture with AFPN (FMA)-fully convolutional one-stage (FCOS)] is proposed. To validate the effectiveness of FMA-FCOS, training, and testing are conducted on the multiscenario vessel dataset collected specifically for this study and the nuScenes dataset. In comparison with methods solely relying on a camera, such as the original FCOS object detection framework and YOLOv9, as well as other fusion methodologies combining camera and radar, the results demonstrate that FMA-FCOS delivers notable advantages, achieving a superior or comparable detection accuracy in the datasets.
引用
收藏
页码:31562 / 31572
页数:11
相关论文
共 42 条
[1]   Emerging Trends in Autonomous Vehicle Perception: Multimodal Fusion for 3D Object Detection [J].
Alaba, Simegnew Yihunie ;
Gurbuz, Ali C. ;
Ball, John E. .
WORLD ELECTRIC VEHICLE JOURNAL, 2024, 15 (01)
[2]   Robust Detection and Tracking Method for Moving Object Based on Radar and Camera Data Fusion [J].
Bai, Jie ;
Li, Sen ;
Huang, Libo ;
Chen, Huanlei .
IEEE SENSORS JOURNAL, 2021, 21 (09) :10761-10774
[3]   LWDNet-A lightweight water-obstacles detection network for unmanned surface vehicles [J].
Cai, Qilie ;
Wang, Qiang ;
Zhang, Yulong ;
He, Zhibo ;
Zhang, Yuhong .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 166
[4]   Spatial Attention Fusion for Obstacle Detection Using MmWave Radar and Vision Sensor [J].
Chang, Shuo ;
Zhang, Yifan ;
Zhang, Fan ;
Zhao, Xiaotong ;
Huang, Sai ;
Feng, Zhiyong ;
Wei, Zhiqing .
SENSORS, 2020, 20 (04)
[5]   Multi-View 3D Object Detection Network for Autonomous Driving [J].
Chen, Xiaozhi ;
Ma, Huimin ;
Wan, Ji ;
Li, Bo ;
Xia, Tian .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6526-6534
[6]   FUTR3D: A Unified Sensor Fusion Framework for 3D Detection [J].
Chen, Xuanyao ;
Zhang, Tianyuan ;
Wang, Yue ;
Wang, Yilun ;
Zhao, Hang .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW, 2023, :172-181
[7]  
Darweesh H, 2017, J ROBOT MECHATRON, V29, P668, DOI 10.20965/jrm.2017.p0668
[8]   CenterNet: Keypoint Triplets for Object Detection [J].
Duan, Kaiwen ;
Bai, Song ;
Xie, Lingxi ;
Qi, Honggang ;
Huang, Qingming ;
Tian, Qi .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :6568-6577
[9]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[10]   Rich feature hierarchies for accurate object detection and semantic segmentation [J].
Girshick, Ross ;
Donahue, Jeff ;
Darrell, Trevor ;
Malik, Jitendra .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :580-587