Point Cloud 3D Object Detection Based on Improved SECOND Algorithm

被引:0
作者
Zhang Ying [1 ]
Jiang Liangliang [1 ]
Zhang Dongbo [1 ]
Duan Wanlin [1 ]
Sun Yue [1 ]
机构
[1] Xiangtan Univ, Coll Automat & Elect Informat, Xiangtan 411105, Hunan, Peoples R China
关键词
autonomous driving; three-dimensional object detection; feature fusion; loss function;
D O I
10.3788/LOP231016
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Rapid identification and precise positioning of surrounding targets are prerequisites and represent the foundation for safe autonomous vehicle driving. A point cloud 3D object detection algorithm based on an improved SECOND algorithm is proposed to address the challenges of inaccurate recognition and positioning in voxel-based point cloud 3D object detection methods. First, an adaptive spatial feature fusion module is introduced into a 2D convolutional backbone network to fuse spatial features of different scales, so as to improve the model's feature expression capability. Second, by fully utilizing the correlation between bounding box parameters, the three-dimensional distance-intersection over union (3D DIoU) is adopted as the bounding box localization regression loss function, thus improving regression task efficiency. Finally, considering both the classification confidence and positioning accuracy of candidate boxes, a new candidate box quality evaluation standard is utilized to obtain smoother regression results. Experimental results on the KITTI test set demonstrate that the 3D detection accuracy of the proposed algorithm is superior to many previous algorithms. Compared with the SECOND benchmark algorithm, the car and cyclist classes improves by 2. 86 and 3. 84 percentage points, respectively, under simple difficulty; 2. 99 and 3. 89 percentage points, respectively, under medium difficulty; and 7. 06 and 4. 27 percentage points, respectively, under difficult difficulty.
引用
收藏
页数:10
相关论文
共 24 条
[1]   Lidar 3D Target Detection Based on Improved PointPillars [J].
Chen Dejiang ;
Yu Wenjun ;
Gao Yongbin .
LASER & OPTOELECTRONICS PROGRESS, 2023, 60 (10)
[2]  
Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
[3]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[4]   DVFENet: Dual-branch voxel feature extraction network for 3D object detection [J].
He, Yunqian ;
Xia, Guihua ;
Luo, Yongkang ;
Su, Li ;
Zhang, Zhi ;
Li, Wanyi ;
Wang, Peng .
NEUROCOMPUTING, 2021, 459 :201-211
[5]  
Ku J, 2018, IEEE INT C INT ROBOT, P5750, DOI 10.1109/IROS.2018.8594049
[6]   PointPillars: Fast Encoders for Object Detection from Point Clouds [J].
Lang, Alex H. ;
Vora, Sourabh ;
Caesar, Holger ;
Zhou, Lubing ;
Yang, Jiong ;
Beijbom, Oscar .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :12689-12697
[7]   3D Object Detection Based on Improved Frustum PointNet [J].
Liu Xunhua ;
Sun Shaoyuan ;
Gu Lipeng ;
Li Xiang .
LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (20)
[8]   AutoShape: Real-Time Shape-Aware Monocular 3D Object Detection [J].
Liu, Zongdai ;
Zhou, Dingfu ;
Lu, Feixiang ;
Fang, Jin ;
Zhang, Liangjun .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :15621-15630
[9]   Frustum PointNets for 3D Object Detection from RGB-D Data [J].
Qi, Charles R. ;
Liu, Wei ;
Wu, Chenxia ;
Su, Hao ;
Guibas, Leonidas J. .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :918-927
[10]  
Qi CR, 2017, Arxiv, DOI [arXiv:1706.02413, 10.48550/arxiv.1706.02413, DOI 10.48550/ARXIV.1706.02413]