FP-RCNN: A Real-Time 3D Target Detection Model based on Multiple Foreground Point Sampling for Autonomous Driving

被引:2
作者
Xu, Guoqing [1 ]
Xu, Xiaolong [2 ]
Gao, Honghao [3 ]
Xiao, Fu [2 ]
机构
[1] Nanjing Univ Posts & Telecommun, Jiangsu Key Lab Big Data Secur & Intelligent Proc, Nanjing, Peoples R China
[2] Nanjing Univ Posts & Telecommun, Sch Comp Sci, Nanjing, Peoples R China
[3] Shanghai Univ, Sch Comp Engn & Sci, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Autonomous driving; Deep learning; 3D target detection; Instance-aware downsampling; Anchor-free; OBJECT DETECTION;
D O I
10.1007/s11036-023-02092-z
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The perception module of autonomous driving needs to maintain high detection accuracy and speed in various weather conditions. Two-dimensional target detection maintains fast detection speed but poor detection accuracy in bad weather, whereas three-dimensional (3D) target detection can still have a good detection effect in bad weather. However, in the current 3D target detection methods, the accuracy of single-stage detection algorithms is insufficient, and the speed of two-stage detection algorithms is slow. Therefore, in this study, we propose a real-time 3D target detection model based on multiple foreground point sampling for autonomous driving, FP-RCNN. FP-RCNN incorporates features from the original points, voxels, and birds-eye view (BEV), uses sparse convolution at the voxel level, performs multiple progressive downsampling to extract features, and maps the obtained features on BEV to obtain BEV features. A three-layer progressive sampling structure was used for key point sampling. The third layer uses instance-aware downsampling to exploit semantic information to ensure that as many foreground points as possible are collected, and the three features are subjected to VSA operations to obtain the final features bound to sampled key points. The second stage divides the proposed box obtained in the first stage, fuses the contextual information of the original points to obtain the final point features, and outputs the confidence box through two fully connected layers. FP-RCNN is tested on the KITTI dataset, and the test results show a 6% improvement in pedestrian detection and a 50% improvement in detection speed compared with a representative two-stage approach.
引用
收藏
页码:369 / 381
页数:13
相关论文
共 50 条
[21]   MSSF: A 4D Radar and Camera Fusion Framework With Multi-Stage Sampling for 3D Object Detection in Autonomous Driving [J].
Liu, Hongsi ;
Liu, Jun ;
Jiang, Guangfeng ;
Jin, Xin .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2025, 26 (06) :8641-8656
[22]   LiDAR-based 3D Object Detection for Autonomous Driving [J].
Li, Zirui .
2022 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, COMPUTER VISION AND MACHINE LEARNING (ICICML), 2022, :507-512
[23]   Real-time accident anticipation for autonomous driving through monocular depth-enhanced 3D modeling [J].
Liao, Haicheng ;
Li, Yongkang ;
Li, Zhenning ;
Bian, Zilin ;
Lee, Jaeyoung ;
Cui, Zhiyong ;
Zhang, Guohui ;
Xu, Chengzhong .
ACCIDENT ANALYSIS AND PREVENTION, 2024, 207
[24]   DSC3D: Deformable Sampling Constraints in Stereo 3D Object Detection for Autonomous Driving [J].
Chen, Jiawei ;
Song, Qi ;
Guo, Wenzhong ;
Huang, Rui .
IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) :2794-2805
[25]   Real-time traffic sign detection system for autonomous driving based on YOLO algorithm [J].
Qiu, Linrun .
2024 CROSS STRAIT RADIO SCIENCE AND WIRELESS TECHNOLOGY CONFERENCE, CSRSWTC 2024, 2024, :351-354
[26]   SpikiLi: A Spiking Simulation of LiDAR based Real-time Object Detection for Autonomous Driving [J].
Mohapatra, Sambit ;
Mesquida, Thomas ;
Hodaei, Mona ;
Yogamani, Senthil ;
Gotzig, Heinrich ;
Maeder, Patrick .
2022 8TH INTERNATIONAL CONFERENCE ON EVENT-BASED CONTROL, COMMUNICATION AND SIGNAL PROCESSING (EBCCSP 2022), 2022,
[27]   LS-YOLO: A Lightweight, Real-Time YOLO-Based Target Detection Algorithm for Autonomous Driving Under Adverse Environmental Conditions [J].
Ju, Cheng ;
Chang, Yuxin ;
Xie, Yuansha ;
Li, Dina .
IEEE ACCESS, 2025, 13 :118147-118162
[28]   Exploring Diversity-Based Active Learning for 3D Object Detection in Autonomous Driving [J].
Lin, Jinpeng ;
Liang, Zhihao ;
Deng, Shengheng ;
Cai, Lile ;
Jiang, Tao ;
Li, Tianrui ;
Jia, Kui ;
Xu, Xun .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (11) :15454-15466
[29]   Transformer-Based Optimized Multimodal Fusion for 3D Object Detection in Autonomous Driving [J].
Alaba, Simegnew Yihunie ;
Ball, John E. .
IEEE ACCESS, 2024, 12 :50165-50176
[30]   Real-time 3D Object Detection from Point Clouds using an RGB-D Camera [J].
Wang, Ya ;
Xu, Shu ;
Zell, Andreas .
ICPRAM: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION APPLICATIONS AND METHODS, 2020, :407-414