Semantic RGB-D SLAM for Rescue Robot Navigation

被引:18
作者
Deng, Wenbang [1 ]
Huang, Kaihong [1 ]
Chen, Xieyuanli [2 ]
Zhou, Zhiqian [1 ]
Shi, Chenghao [1 ]
Guo, Ruibin [1 ]
Zhang, Hui [1 ]
机构
[1] Natl Univ Def Technol, Coll Intelligence Sci & Technol, Robot Res Ctr, Changsha 410073, Peoples R China
[2] Univ Bonn, Photogrammetry & Robot Lab, D-53115 Bonn, Germany
基金
中国国家自然科学基金;
关键词
Semantics; Simultaneous localization and mapping; Image segmentation; Cameras; Rescue robots; Navigation; Deep learning; path planning; RoboCup; rescue robot; semantic SLAM; LARGE-SCALE; MAP;
D O I
10.1109/ACCESS.2020.3031867
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we propose a semantic simultaneous localization and mapping (SLAM) framework for rescue robots, and report its use in navigation tasks. Our framework can generate not only geometric maps in the form of dense point-clouds but also corresponding point-wise semantic labels generated by a semantic segmentation convolutional neural network (CNN). The semantic segmentation CNN is trained using our RGB-D dataset of the RoboCup Rescue-Robot-League (RRL) competition environment. With the help of semantic information, the rescue robot can identify different types of terrains in a complex environment, so as to avoid specific obstacles or to choose routes with better traversability. To reduce the segmentation noise, our approach utilizes depth images to perform filtering on the segmentation results of each frame. The overall semantic map is then further improved in the point-cloud voxels. By accumulating results of multiple frames in the voxels, semantic maps with consistent semantic labels are obtained. To show the advantage of having a semantic map of the environment, we report a case study of how the semantic map can be utilized in a navigation task to reduce the arrival time while ensuring safety. The experimental result shows that our semantic SLAM framework is capable of generating a dense semantic map for the complex RRL competition environment, with which the arrival time of the navigation time is effectively reduced.
引用
收藏
页码:221320 / 221329
页数:10
相关论文
共 42 条
[1]  
[Anonymous], 2016, ARXIV161104144
[2]  
Antonello M, 2018, IEEE INT CONF ROBOT, P1855
[3]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[4]  
Bowman Sean L., 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA), P1722, DOI 10.1109/ICRA.2017.7989203
[5]  
Chen L., 2014, INT C LEARN REPR, V4, P357
[6]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[7]  
Chen X., 2020, ROB SCI SYST 16
[8]  
Chen XYL, 2019, IEEE INT C INT ROBOT, P4530, DOI 10.1109/IROS40897.2019.8967704
[9]   The Cityscapes Dataset for Semantic Urban Scene Understanding [J].
Cordts, Marius ;
Omran, Mohamed ;
Ramos, Sebastian ;
Rehfeld, Timo ;
Enzweiler, Markus ;
Benenson, Rodrigo ;
Franke, Uwe ;
Roth, Stefan ;
Schiele, Bernt .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3213-3223
[10]   Real-Time Semantic Plane Reconstruction on a Monocular Drone Using Sparse Fusion [J].
Dang, Yuanjie ;
Chen, Peng ;
Liang, Ronghua ;
Huang, Chong ;
Tang, Yuesheng ;
Yu, Tianwei ;
Yang, Xin ;
Cheng, Kwang-Ting .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (08) :7383-7391