Collaborative 3D Object Detection for Autonomous Vehicles via Learnable Communications

被引:5
作者
Wang, J. [1 ]
Zeng, Y. [2 ]
Gong, Y. [1 ]
机构
[1] Southern Univ Sci & Technol, Dept Elect & Elect Engn, Shenzhen 518055, Peoples R China
[2] Southern Univ Sci & Technol, Res Inst Trustworthy Autonomous Syst, Shenzhen 518055, Peoples R China
基金
中国国家自然科学基金;
关键词
Three-dimensional displays; Object detection; Point cloud compression; Feature extraction; Collaboration; Bandwidth; Autonomous vehicles; Collaborative perception; learnable communications; 3D object detection; autonomous driving;
D O I
10.1109/TITS.2023.3272027
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
3D object detection from LiDAR point cloud is a challenging task in autonomous driving systems. Collaborative perception can incorporate information from spatially diverse sensors and provide significant benefits for accurate 3D object detection from point clouds. In this work, we consider that the autonomous vehicle uses local point cloud data and combines information from neighboring infrastructures through wireless links for cooperative 3D object detection. However, information sharing among vehicles and infrastructures in predefined communication schemes may result in communication congestion and/or bring limited performance improvement. To this end, we propose a novel collaborative 3D object detection framework using an encoder-decoder network architecture and an attention-based learnable communications scheme. It consists of three components: a feature encoder network that maps point clouds into feature maps; an attention-based communication module that propagates compact and fine-grained query feature maps from the vehicle to support infrastructures, and optimizes attention weights between query and key to refine support feature maps; a region proposal network that fuses local feature maps and weighted support feature maps for 3D object detection. We evaluate the performance of the proposed framework on CARLA-3D, a new dataset that we synthesized using CARLA for 3D cooperative object detection. Experimental results and bandwidth consumption analysis show that the proposed collaborative 3D object detection framework achieves a better detection performance and communication bandwidth trade-off than five baseline 3D object detection models under different detection difficulties.
引用
收藏
页码:9804 / 9816
页数:13
相关论文
共 43 条
[1]   Cooperative Perception for 3D Object Detection in Driving Scenarios Using Infrastructure Sensors [J].
Arnold, Eduardo ;
Dianati, Mehrdad ;
de Temple, Robert ;
Fallah, Saber .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (03) :1852-1864
[2]   A Survey on 3D Object Detection Methods for Autonomous Driving Applications [J].
Arnold, Eduardo ;
Al-Jarrah, Omar Y. ;
Dianati, Mehrdad ;
Fallah, Saber ;
Oxtoby, David ;
Mouzakitis, Alex .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2019, 20 (10) :3782-3795
[3]  
Beltrán J, 2018, IEEE INT C INTELL TR, P3517, DOI 10.1109/ITSC.2018.8569311
[4]   Vehicle-to-vehicle wireless communication protocols for enhancing highway traffic safety [J].
Biswas, S ;
Tatchikou, R ;
Dion, F .
IEEE COMMUNICATIONS MAGAZINE, 2006, 44 (01) :74-82
[5]   F-Cooper: Feature based Cooperative Perception for Autonomous Vehicle Edge Computing System Using 3D Point Clouds [J].
Chen, Qi ;
Ma, Xu ;
Tang, Sihai ;
Guo, Jingda ;
Yang, Qing ;
Fu, Song .
SEC'19: PROCEEDINGS OF THE 4TH ACM/IEEE SYMPOSIUM ON EDGE COMPUTING, 2019, :88-100
[6]   Cooper: Cooperative Perception for Connected Autonomous Vehicles based on 3D Point Clouds [J].
Chen, Qi ;
Tang, Sihai ;
Yang, Qing ;
Fu, Song .
2019 39TH IEEE INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2019), 2019, :514-524
[7]   Shape Completion using 3D-Encoder-Predictor CNNs and Shape Synthesis [J].
Dai, Angela ;
Qi, Charles Ruizhongtai ;
Niessner, Matthias .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :6545-6554
[8]  
Das A, 2019, P MACHINE LEARNING R, V97
[9]  
Dosovitskiy A., 2017, P 1 ANN C ROB LEARN, P1
[10]  
Dumoulin V, 2018, Arxiv, DOI [arXiv:1603.07285, 10.48550/arXiv.1603.07285]