Autonomous Driving 3D Object Detection Based on Cascade YOLOv7

被引:0
|
作者
Zhao D. [1 ]
Zhao S. [1 ]
机构
[1] School of Mechatronics and Vehicle Engineering, Chongqing Jiaotong University, Chongqing
来源
Qiche Gongcheng/Automotive Engineering | 2023年 / 45卷 / 07期
关键词
3D object detection; autonomous driving; F-PointNet; multi-sensor information fusion; YOLOv7;
D O I
10.19562/j.chinasae.qcgc.2023.07.002
中图分类号
学科分类号
摘要
For the problems of incomplete feature information and excessive point cloud search volume in 3D object detection methods based on image and original point cloud,based on Frustum PointNet structure,a 3D object detection algorithm based on cascade YOLOv7 is proposed by fusing RGB image and point cloud data of autonomous driving surrounding scenes. Firstly,a frustum estimation model based on YOLOv7 is constructed to longitudinally expand the RGB image RoI into 3D space. Then the object point cloud and background point cloud in the frustum are segmented by PointNet ++. Finally,the natural position relationship between objects is explained by using the non-modal 3D box estimation network to output the length,width,height,heading et al. of objects. The test results and ablation experiments on the KITTI public dataset show that compared with the benchmark network,the inference time of cascade YOLOv7 model is shortened by 40 ms∙frame-1,with the mean average precision of detection at the moderate,difficulty level increased by 8.77%,9.81%,respectively. © 2023 SAE-China. All rights reserved.
引用
收藏
页码:1112 / 1122
页数:10
相关论文
共 29 条
  • [1] LI B, OUYANG W, SHENG L, Et al., Gs3D:an efficient 3D object detection framework for autonomous driving[C], Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 1019-1028, (2019)
  • [2] GIRSHICK R., Fast R-CNN[C], Proceedings of the IEEE International Conference on Computer Vision, pp. 1440-1448, (2015)
  • [3] LI P, CHEN X, SHEN S., Stereo R-CNN based 3D object detection for autonomous driving[C], Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 7644-7652, (2019)
  • [4] HE K, GKIOXARI G, DOLLAR P, Et al., Mask R-CNN[C], Proceedings of the IEEE International Conference on Computer Vision, pp. 2961-2969, (2017)
  • [5] BEWLEY A, SUN P, MENSINK T, Et al., Range conditioned dilated convolutions for scale invariant 3D object detection[J], (2020)
  • [6] LI Q, LI Y C,, ZHANG Y L,, Et al., A single-stage point cloud 3D object detection method using sparse 3D convolution[J], Journal of Xi’an Jiaotong University, 56, 9, pp. 112-122, (2022)
  • [7] CAESAR H, BANKITI V, LANG A H, Et al., Nuscenes:a multimodal dataset for autonomous driving[C], Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11621-11631, (2020)
  • [8] HU J S K, WASLANDER S L., Point density-aware voxels for lidar 3D object detection[C], Proceedings of the IEEE/ CVF Conference on Computer Vision and Pattern Recognition, pp. 8469-8478, (2022)
  • [9] ZHANG M F, WU Y F,, WANG L, Et al., Pyramid-feature-fusion-based two stage vehicle detection via 3D point cloud[J], Journal of Transportation Systems Engineering and Information Technology, 22, 5, pp. 107-116, (2022)
  • [10] SHI S, WANG X, LI H., PointRCNN:3D object proposal generation and detection from point cloud[C], Proceedings of the IEEE/ CVF Conference on Computer Vision and Pattern Recognition, pp. 770-779, (2019)