SO-Net: Joint Semantic Segmentation and Obstacle Detection Using Deep Fusion of Monocular Camera and Radar

被引:15
作者
John, V [1 ]
Nithilan, M. K. [1 ]
Mita, S. [1 ]
Tehrani, H. [1 ,2 ]
Sudheesh, R. S. [2 ]
Lalu, P. P. [2 ]
机构
[1] Toyota Technol Inst, Res Ctr Smart Vehicles, Nagoya, Aichi, Japan
[2] Govt Engn Coll, Nodal Ctr Robot & Artificial Intelligence, Trichur, India
来源
IMAGE AND VIDEO TECHNOLOGY, PSIVT 2019 INTERNATIONAL WORKSHOPS | 2020年 / 11994卷
关键词
Joint learning; Sensor fusion; Radar; Monocular camera;
D O I
10.1007/978-3-030-39770-8_11
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Vision-based semantic segmentation and obstacle detection are important perception tasks for autonomous driving. Vision-based semantic segmentation and obstacle detection are performed using separate frameworks resulting in increased computational complexity. Vision-based perception using deep learning reports state-of-the-art accuracy, but the performance is susceptible to variations in the environment. In this paper, we propose a radar and vision-based deep learning perception framework termed as the SO-Net to address the limitations of vision-based perception. The SO-Net also integrates the semantic segmentation and object detection within a single framework. The proposed SO-Net contains two input branches and two output branches. The SO-Net input branches correspond to vision and radar feature extraction branches. The output branches correspond to object detection and semantic segmentation branches. The performance of the proposed framework is validated on the Nuscenes public dataset. The results show that the SO-Net improves the accuracy of the vision-only-based perception tasks. The SO-Net also reports reduced computational complexity compared to separate semantic segmentation and object detection frameworks.
引用
收藏
页码:138 / 148
页数:11
相关论文
共 20 条
[1]  
Bombini L., 2006, INT WORKSH INT TRANS, P65
[2]  
Caesar H, 2019, P IEEECVF C COMPUTER
[3]  
Chadwick S, 2019, IEEE INT CONF ROBOT, P8311, DOI [10.1109/icra.2019.8794312, 10.1109/ICRA.2019.8794312]
[4]   Depth-based target segmentation for intelligent vehicles: Fusion of radar and binocular stereo [J].
Fang, YJ ;
Masaki, I ;
Horn, B .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2002, 3 (03) :196-202
[5]  
Gaisser F, 2017, PROCEEDINGS OF THE FIFTEENTH IAPR INTERNATIONAL CONFERENCE ON MACHINE VISION APPLICATIONS - MVA2017, P101, DOI 10.23919/MVA.2017.7986800
[6]  
Garcia F, 2012, 2012 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), P494, DOI 10.1109/IVS.2012.6232199
[7]   Vehicle Detection and Tracking in Car Video Based on Motion Model [J].
Jazayeri, Amirali ;
Cai, Hongyuan ;
Zheng, Jiang Yu ;
Tuceryan, Mihran .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2011, 12 (02) :583-595
[8]   RVNet: Deep Sensor Fusion of Monocular Camera and Radar for Image-Based Obstacle Detection in Challenging Environments [J].
John, Vijay ;
Mita, Seiichi .
IMAGE AND VIDEO TECHNOLOGY (PSIVT 2019), 2019, 11854 :351-364
[9]  
John V, 2018, INT C PATT RECOG, P189, DOI 10.1109/ICPR.2018.8546108
[10]   An obstacle detection method by fusion of radar and motion stereo [J].
Kato, T ;
Ninomiya, Y ;
Masaki, I .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2002, 3 (03) :182-188