Dynamic Semantic SLAM Based on Panoramic Camera and LiDAR Fusion for Autonomous Driving

被引:0
作者
Li, Feiya [1 ]
Fu, Chunyun [1 ]
Wang, Jianwen [2 ]
Sun, Dongye [3 ]
机构
[1] Chongqing Univ, Coll Mech & Vehicle Engn, Chongqing 400044, Peoples R China
[2] Chongqing Changan Automobile Co Ltd, State Key Lab Intelligent Vehicle Safety Technol, Chongqing 400023, Peoples R China
[3] Chongqing Univ, State Key Lab Mech Transmiss Adv Equipment, Chongqing 400044, Peoples R China
关键词
Laser radar; Simultaneous localization and mapping; Point cloud compression; Semantics; Optimization; Location awareness; Heuristic algorithms; Pose estimation; Feature extraction; Cameras; Autonomous driving; semantic SLAM; camera and LiDAR fusion; dynamic environment; traffic engineering; pose estimation;
D O I
10.1109/TITS.2025.3550561
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
This paper proposes a semantic SLAM methodology for autonomous driving, integrating panoramic camera and LiDAR fusion to address pose estimation inaccuracies and lack of semantic information in dynamic environments. On the LiDAR side, panoramic semantic segmentation results are projected onto point clouds and combined with clustering for instance segmentation, classifying points into unknown-motion landmarks, static landmarks, and ground points. Precise pose estimation is then achieved using purely static and semi-static landmarks. The visual component incorporates data augmentation and instance segmentation of front-view images, leveraging SuperPoint for feature extraction. 3D dynamic landmarks are projected onto segmentation results to generate dynamic-static masks, enabling the elimination of feature points within dynamic regions. A robust visual odometry system is implemented through the ORB-SLAM2 tracking framework, enhanced by SuperGlue feature association within a multi-level tracking mechanism. At the fusion level, a decision-layer fusion algorithm integrates both odometry sources. Furthermore, pose optimization is achieved through dual loop closure detection with inputs from both LiDAR and visual sources. The proposed system was validated using the CQU01-CQU04 datasets collected on a UGV, with high-precision GNSS serving as the ground truth. Experimental results demonstrate that the system achieves mean of average error of estimated positions, standard deviation of estimated position errors, and position error at the trajectory end of 0.446m, 0.236m, and 0.775m, respectively. With loop closure enabled, the performance metrics showed enhancement to 0.320m, 0.181m, and 0.617m. This research demonstrates that our method improves localization accuracy while constructing semantic map comprising purely static elements.
引用
收藏
页数:14
相关论文
共 43 条
[1]  
Arini Parhusip H., 2024, IAES Int. J. Artif. Intell. (IJ-AI), V13, P3291
[2]  
Cao K., 2023, P IEEE INT C ROB BIO, P1
[3]   DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs [J].
Chen, Liang-Chieh ;
Papandreou, George ;
Kokkinos, Iasonas ;
Murphy, Kevin ;
Yuille, Alan L. .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (04) :834-848
[4]   SLAM Overview: From Single Sensor to Heterogeneous Fusion [J].
Chen, Weifeng ;
Zhou, Chengjun ;
Shang, Guangtao ;
Wang, Xiyang ;
Li, Zhenxiong ;
Xu, Chonghui ;
Hu, Kai .
REMOTE SENSING, 2022, 14 (23)
[5]   A review of visual SLAM methods for autonomous driving vehicles [J].
Cheng, Jun ;
Zhang, Liyan ;
Chen, Qihong ;
Hu, Xinrong ;
Cai, Jingcao .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 114
[6]   LinK3D: Linear Keypoints Representation for 3D LiDAR Point Cloud [J].
Cui, Yunge ;
Zhang, Yinlong ;
Dong, Jiahua ;
Sun, Haibo ;
Chen, Xieyuanli ;
Zhu, Feng .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (03) :2128-2135
[7]   BoW3D: Bag of Words for Real-Time Loop Closing in 3D LiDAR SLAM [J].
Cui, Yunge ;
Chen, Xieyuanli ;
Zhang, Yinlong ;
Dong, Jiahua ;
Wu, Qingxiao ;
Zhu, Feng .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (05) :2828-2835
[8]   SuperPoint: Self-Supervised Interest Point Detection and Description [J].
DeTone, Daniel ;
Malisiewicz, Tomasz ;
Rabinovich, Andrew .
PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, :337-349
[9]  
Dong X, 2011, IEEE INT CON MULTI
[10]  
Graeter J, 2018, IEEE INT C INT ROBOT, P7872, DOI 10.1109/IROS.2018.8594394