Real-Time Metric-Semantic Mapping for Autonomous Navigation in Outdoor Environments

被引:4
作者
Jiao, Jianhao [1 ]
Geng, Ruoyu [2 ]
Li, Yuanhang [3 ]
Xin, Ren [2 ]
Yang, Bowen [3 ]
Wu, Jin [3 ]
Wang, Lujia [4 ]
Liu, Ming [2 ]
Fan, Rui [5 ,6 ]
Kanoulas, Dimitrios [1 ]
机构
[1] UCL, Dept Comp Sci, London WC1E 6BT, England
[2] Hong Kong Univ Sci & Technol Guangzhou, Robot & Autonomous Syst, Guangzhou 511400, Guangdong, Peoples R China
[3] Hong Kong Univ Sci & Technol, Dept Elect & Comp Engn, Hong Kong, Peoples R China
[4] Hong Kong Univ Sci & Technol Guangzhou, Guangzhou 511400, Guangdong, Peoples R China
[5] Tongji Univ, Shanghai Res Inst Intelligent Autonomous Syst, Coll Elect & Informat Engn, Shanghai 201804, Peoples R China
[6] Tongji Univ, Frontiers Sci Ctr Intelligent Autonomous Syst, Shanghai 201804, Peoples R China
基金
中国国家自然科学基金; 英国科研创新办公室;
关键词
Semantics; Navigation; Robots; Robot sensing systems; Sensors; Real-time systems; Location awareness; Autonomous driving; mapping; navigation;
D O I
10.1109/TASE.2024.3429280
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The creation of a metric-semantic map, which encodes human-prior knowledge, represents a high-level abstraction of environments. However, constructing such a map poses challenges related to the fusion of multi-modal sensor data, the attainment of real-time mapping performance, and the preservation of structural and semantic information consistency. In this paper, we introduce an online metric-semantic mapping system that utilizes LiDAR-Visual-Inertial sensing to generate a global metric-semantic mesh map of large-scale outdoor environments. Leveraging GPU acceleration, our mapping process achieves exceptional speed, with frame processing taking less than 7ms, regardless of scenario scale. Furthermore, we seamlessly integrate the resultant map into a real-world navigation system, enabling metric-semantic-based terrain assessment and autonomous point-to-point navigation within a campus environment. Through extensive experiments conducted on both publicly available and self-collected datasets comprising 24 sequences, we demonstrate the effectiveness of our mapping and navigation methodologies. Note to Practitioners-This paper tackles the challenge of autonomous navigation for mobile robots in complex, unstructured environments with rich semantic elements. Traditional navigation relies on geometric analysis and manual annotations, struggling to differentiate similar structures like roads and sidewalks. We propose an online mapping system that creates a global metric-semantic mesh map for large-scale outdoor environments, utilizing GPU acceleration for speed and overcoming the limitations of existing real-time semantic mapping methods, which are generally confined to indoor settings. Our map integrates into a real-world navigation system, proven effective in localization and terrain assessment through experiments with both public and proprietary datasets. Future work will focus on integrating kernel-based methods to improve the map's semantic accuracy.
引用
收藏
页码:5729 / 5740
页数:12
相关论文
共 48 条
[1]   A Benchmark for LiDAR-based Panoptic Segmentation based on KITTI [J].
Behley, Jens ;
Milioto, Andres ;
Stachniss, Cyrill .
2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, :13596-13603
[2]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[3]   Moving Object Segmentation in 3D LiDAR Data: A Learning-Based Approach Exploiting Sequential Data [J].
Chen, Xieyuanli ;
Li, Shijie ;
Mersch, Benedikt ;
Wiesmann, Louis ;
Gall, Jurgen ;
Behley, Jens ;
Stachniss, Cyrill .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (04) :6529-6536
[4]   The Cityscapes Dataset for Semantic Urban Scene Understanding [J].
Cordts, Marius ;
Omran, Mohamed ;
Ramos, Sebastian ;
Rehfeld, Timo ;
Enzweiler, Markus ;
Benenson, Rodrigo ;
Franke, Uwe ;
Roth, Stefan ;
Schiele, Bernt .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3213-3223
[5]   Learning-Aided 3-D Occupancy Mapping With Bayesian Generalized Kernel Inference [J].
Doherty, Kevin ;
Shan, Tixiao ;
Wang, Jinkun ;
Englot, Brendan .
IEEE TRANSACTIONS ON ROBOTICS, 2019, 35 (04) :953-966
[6]  
Furgale P, 2013, IEEE INT C INT ROBOT, P1280, DOI 10.1109/IROS.2013.6696514
[7]   Visual Teach and Repeat for Long-Range Rover Autonomy [J].
Furgale, Paul ;
Barfoot, Timothy D. .
JOURNAL OF FIELD ROBOTICS, 2010, 27 (05) :534-560
[8]   Multitask Learning for Scalable and Dense Multilayer Bayesian Map Inference [J].
Gan, Lu ;
Kim, Youngji ;
Grizzle, Jessy W. ;
Walls, Jeffrey M. ;
Kim, Ayoung ;
Eustice, Ryan M. ;
Ghaffari, Maani .
IEEE TRANSACTIONS ON ROBOTICS, 2023, 39 (01) :699-717
[9]   Bayesian Spatial Kernel Smoothing for Scalable Dense Semantic Mapping [J].
Gan, Lu ;
Zhang, Ray ;
Grizzle, Jessy W. ;
Eustice, Ryan M. ;
Ghaffari, Maani .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2020, 5 (02) :790-797
[10]   Volumetric Instance-Aware Semantic Mapping and 3D Object Discovery [J].
Grinvald, Margarita ;
Furrer, Fadri ;
Novkovic, Tonci ;
Chung, Jen Jen ;
Cadena, Cesar ;
Siegwart, Roland ;
Nieto, Juan .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (03) :3037-3044