FP-RCNN: A Real-Time 3D Target Detection Model based on Multiple Foreground Point Sampling for Autonomous Driving

被引:2
|
作者
Xu, Guoqing [1 ]
Xu, Xiaolong [2 ]
Gao, Honghao [3 ]
Xiao, Fu [2 ]
机构
[1] Nanjing Univ Posts & Telecommun, Jiangsu Key Lab Big Data Secur & Intelligent Proc, Nanjing, Peoples R China
[2] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Nanjing, Peoples R China
[3] Shanghai Univ, Sch Comp Engn & Sci, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Autonomous driving; Deep learning; 3D target detection; Instance-aware downsampling; Anchor-free; OBJECT DETECTION;
D O I
10.1007/s11036-023-02092-z
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The perception module of autonomous driving needs to maintain high detection accuracy and speed in various weather conditions. Two-dimensional target detection maintains fast detection speed but poor detection accuracy in bad weather, whereas three-dimensional (3D) target detection can still have a good detection effect in bad weather. However, in the current 3D target detection methods, the accuracy of single-stage detection algorithms is insufficient, and the speed of two-stage detection algorithms is slow. Therefore, in this study, we propose a real-time 3D target detection model based on multiple foreground point sampling for autonomous driving, FP-RCNN. FP-RCNN incorporates features from the original points, voxels, and birds-eye view (BEV), uses sparse convolution at the voxel level, performs multiple progressive downsampling to extract features, and maps the obtained features on BEV to obtain BEV features. A three-layer progressive sampling structure was used for key point sampling. The third layer uses instance-aware downsampling to exploit semantic information to ensure that as many foreground points as possible are collected, and the three features are subjected to VSA operations to obtain the final features bound to sampled key points. The second stage divides the proposed box obtained in the first stage, fuses the contextual information of the original points to obtain the final point features, and outputs the confidence box through two fully connected layers. FP-RCNN is tested on the KITTI dataset, and the test results show a 6% improvement in pedestrian detection and a 50% improvement in detection speed compared with a representative two-stage approach.
引用
收藏
页码:369 / 381
页数:13
相关论文
共 50 条
  • [1] FP-RCNN: A Real-Time 3D Target Detection Model based on Multiple Foreground Point Sampling for Autonomous Driving
    Guoqing Xu
    Xiaolong Xu
    Honghao Gao
    Fu Xiao
    Mobile Networks and Applications, 2023, 28 : 369 - 381
  • [2] Real-Time Semantic Segmentation of 3D Point Cloud for Autonomous Driving
    Kang, Dongwan
    Wong, Anthony
    Lee, Banghyon
    Kim, Jungha
    ELECTRONICS, 2021, 10 (16)
  • [3] A RGB-D Based Real-Time Multiple Object Detection and Ranging System for Autonomous Driving
    Yang, Jiachen
    Wang, Chenguang
    Wang, Huihui
    Li, Qiang
    IEEE SENSORS JOURNAL, 2020, 20 (20) : 11959 - 11966
  • [4] BRTPillar: boosting real-time 3D object detection based point cloud and RGB image fusion in autonomous driving
    Zhang, Zhitian
    Zhao, Hongdong
    Zhao, Yazhou
    Chen, Dan
    Zhang, Ke
    Li, Yanqi
    INTERNATIONAL JOURNAL OF INTELLIGENT COMPUTING AND CYBERNETICS, 2025, 18 (01) : 217 - 235
  • [5] Real-Time 3D Object Detection From Point Cloud Through Foreground Segmentation
    Wang, Bo
    Zhu, Ming
    Lu, Ying
    Wang, Jiarong
    Gao, Wen
    Wei, Hua
    IEEE ACCESS, 2021, 9 : 84886 - 84898
  • [6] Deep Learning Based, Real-Time Object Detection for Autonomous Driving
    Akyol, Gamze
    Kantarci, Alperen
    Celik, Ali Eren
    Ak, Abdullah Cihan
    2020 28TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2020,
  • [7] RT3D: Real-Time 3-D Vehicle Detection in LiDAR Point Cloud for Autonomous Driving
    Zeng, Yiming
    Hu, Yu
    Liu, Shice
    Ye, Jing
    Han, Yinhe
    Li, Xiaowei
    Sun, Ninghui
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 3434 - 3440
  • [8] A Decision Fusion Model for 3D Detection of Autonomous Driving
    Ye, Zhen
    Xue, Jianru
    Fang, Jianwu
    Dou, Jian
    Pan, Yuxin
    2018 CHINESE AUTOMATION CONGRESS (CAC), 2018, : 3773 - 3777
  • [9] Real-time rail recognition based on 3D point clouds
    Yu, Xinyi
    He, Weiqi
    Qian, Xuecheng
    Yang, Yang
    Zhang, Tingting
    Ou, Linlin
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2022, 33 (10)
  • [10] CrossFusion net: Deep 3D object detection based on RGB images and point clouds in autonomous driving
    Hong, Dza-Shiang
    Chen, Hung-Hao
    Hsiao, Pei-Yung
    Fu, Li-Chen
    Siao, Siang-Min
    IMAGE AND VISION COMPUTING, 2020, 100