Vehicle-Road Collaborative Perception Method Based on Dual-Stream Feature Extraction

被引:0
作者
NlU, Guochen [1 ]
Sun, Xiangyu [1 ]
Yuan, Zhengyan [1 ]
机构
[1] Robotics Institute, Civil Aviation University of China, Tianjin
来源
Shanghai Jiaotong Daxue Xuebao/Journal of Shanghai Jiaotong University | 2024年 / 58卷 / 11期
关键词
3D object detection; autonomous driving; collaborative perception; feature extraction; information sharing and fusion;
D O I
10.16183/j.cnki.jsjtu.2024.239
中图分类号
学科分类号
摘要
To solve the problem of inadequate perception of autonomous driving in occlusion and over-the-horizon scenarios, a vehicle-road collaborative perception method based on a dual-stream feature extraction network is proposed to enhance the 3D object detection capabilities of traffic participants. Feature extraction networks for roadside and vehicle-side scenes are tailored based on respective characteristics. Since roadside has rich and sufficient sensing data and computational resources, the Transformer structure is used to extract more sophisticated and advanced feature representations. Due to limited computational capability and high real-time demands of autonomous vehicles, partial convolution (PConv) is employed to enhance computing efficiency, and the Mamba-VSS module is introduced for efficient perception in complex environments. Collaborative perception between vehicle-side and roadside is accomplished through the selective sharing and fusion of critical perceptual information guided by confidence maps. By training and testing on DAIR-V2X dataset, the model size of vehicle-side feature extraction network is obtained to be 8. 1 MB, and the IoU thresholds of 0. 5 and 0. 7 correspond to the average accuracy indexes of 67. 67% and 53. 74%. The experiment verifies the advantages of this method in detection accuracy and model size, and provides a lower-configuration detection scheme for vehicle-road collaboration. © 2024 Shanghai Jiaotong University. All rights reserved.
引用
收藏
页码:1826 / 1834
页数:8
相关论文
共 22 条
[1]  
YI Xiaoymg, RUI Yikang, RAN Bin, Vehicle-infrastructure cooperative sensing: Progress and prospect, Strategic Study of CAE, 26, 1, pp. 178-189, (2024)
[2]  
ARNOLD E, DIANATI M, DE TEMPLE R, Et al., Cooperative perception for 3D object detection in driving scenarios using infrastructure sensors, IEEE Transactions on Intelligent Transportation Systems, 23, 3, pp. 1852-1864, (2022)
[3]  
ZHANG Yi, YAO Danya, LI Li, Et al., Technologies and applications for intelligent vehicle-infrastructure cooperation systems, Journal of Transportation Systems Engineering and Information Technology, 21, 5, pp. 40-51, (2021)
[4]  
DOSOVITSKIY A, ROS G, CODER VILLA F, Et al., CARLA: An open urban driving simulator, 1st Conference on Robot Learning, (2017)
[5]  
CHEN Q, TANG S H, YANG Q, Et al., Cooper: Cooperative perception for connected autonomous ve-hicles based on 3D point clouds, 2019 IEEE 39th International Conference on Distributed Computing Systems, pp. 514-524, (2019)
[6]  
CHEN Q., F-cooper: Feature based cooperative per-ception for autonomous vehicle edge computing sys-tem using 3D point clouds
[7]  
GUO J D, CARRILLO D, TANG S H, Et al., CoFF: Cooperative spatial feature fusion for 3-D object de-tection on autonomous vehicles, IEEE Internet of Things Journal, 8, 14, pp. 11078-11087, (2021)
[8]  
HU Y, FANG S, LEI Z, Et al., Where2comm
[9]  
Com-munication-efficient collaborative perception via spa-tial confidence maps, 36th Corference on Neural Information Processing Systems, pp. 4874-4886, (2022)
[10]  
LIU Y C, TIAN J J, GLASER N, Et al., When2coiri: Multi-agent perception via communication graph grouping, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4105-4114, (2020)