Multi-Modality Fusion Perception and Computing in Autonomous Driving

被引:0
作者
Zhang Y. [1 ]
Zhang S. [1 ]
Zhang Y. [1 ]
Ji J. [1 ]
Duan Y. [1 ]
Huang Y. [1 ]
Peng J. [1 ]
Zhang Y. [1 ]
机构
[1] School of Computer Science and Technology, University of Science and Technology of China, Hefei
来源
Jisuanji Yanjiu yu Fazhan/Computer Research and Development | 2020年 / 57卷 / 09期
关键词
Autonomous driving; Computational optimization; Edge computing; Fusion; Multi-modality; Perception;
D O I
10.7544/issn1000-1239.2020.20200255
中图分类号
学科分类号
摘要
The goal of autonomous driving is to provide a safe, comfortable and efficient driving environment for people. In order to have wide-spread deployment of autonomous driving systems, we need to process the sensory data from multiple streams in a timely and accurate fashion. The challenges that arise are thus two-fold: leveraging the multiple sensors that are available on autonomous vehicles to boost the perception accuracy; jointly optimizing perception models and the underlying computing models to meet the real-time requirements. To address these challenges, this paper surveys the latest research on sensing and edge computing for autonomous driving and presents our own autonomous driving system, Sonic. Specifically, we propose a multi-modality perception model, ImageFusion, that combines the lidar data and camera data for 3D object detection, and a computational optimization framework, MPInfer. © 2020, Science Press. All right reserved.
引用
收藏
页码:1781 / 1799
页数:18
相关论文
共 62 条
[1]  
Kroger F., Automated Driving in Its Social, Historical and Cultural Contexts, (2016)
[2]  
Geiger A, Lenz P, Stiller C, Et al., Vision meets robotics: The KITTI dataset, International Journal of Robotics Research, 32, 11, pp. 1231-1237, (2013)
[3]  
Caesar H, Bankiti V, Lang A H, Et al., nuScenes: A multimodal dataset for autonomous driving
[4]  
Chang Mingfang, Lambert J, Sangkloy P, Et al., Argoverse: 3D tracking and forecasting with rich maps, Proc of the IEEE Computer Society Conf on Computer Vision and Pattern Recognition, pp. 8740-8749, (2019)
[5]  
Yurtsever E, Lambert J, Carballo A, Et al., A survey of autonomous driving: Common practices and emerging technologies
[6]  
Chi Lu, Mu Yadong, Learning end-to-end autonomous steering model from spatial and temporal visual cues, Proc of the Workshop on Visual Analysis in Smart and Connected Communities, co-located with MM 2017(VSCC 2017), pp. 9-16, (2017)
[7]  
Behere S, Torngren M., A functional architecture for autonomous driving, Proc of the 2015 ACM Workshop on Automotive Software Architecture, Part of CompArch 2015(WASA 2015), pp. 3-10, (2015)
[8]  
Weber M, Wolf P, Zollner J M., DeepTLR: A single deep convolutional network for detection and classification of traffic lights, Proc of IEEE Intelligent Vehicles Symp, pp. 342-348, (2016)
[9]  
Ranft B, Stiller C., The role of machine vision for intelligent vehicles, IEEE Transactions on Intelligent Vehicles, 1, 1, pp. 8-19, (2016)
[10]  
Lang A H, Vora S, Caesar H, Et al., Pointpillars: Fast encoders for object detection from point clouds, Proc of the IEEE Computer Society Conf on Computer Vision and Pattern Recognition, pp. 12689-12697, (2019)