An efficient point cloud place recognition approach based on transformer in dynamic environment

被引:8
作者
Li, Qipeng [1 ]
Zhuang, Yuan [1 ,2 ]
Huai, Jianzhu [1 ]
Chen, Yiwen [1 ]
Yilmaz, Alper [3 ]
机构
[1] Wuhan Univ, State Key Lab Informat Engn Surveying Mapping & Re, Wuhan 430079, Hubei, Peoples R China
[2] Wuhan Univ, Shenzhen Res Inst, Shenzhen 518057, Guangdong, Peoples R China
[3] Ohio State Univ, Dept Civil Environm & Geodet Engn, 2070 Neil Ave, Columbus, OH 43210 USA
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Place recognition; 3D scene graph representation; Moving object segmentation; Transformer; Global localization; 3D point cloud; CONTEXT; REGISTRATION; LIDAR;
D O I
10.1016/j.isprsjprs.2023.11.013
中图分类号
P9 [自然地理学];
学科分类号
0705 ; 070501 ;
摘要
Place recognition retrieves scenes from a prior map and identifies previously visited locations. It is essential to autonomous-driving and robotics in terms of long-term navigation and global localization. Although the semantic graph has made great progress in 3D place recognition, its construction is greatly affected by the complex dynamic environment. In response, this paper designs a scene graph transformation network to solve place recognition in a dynamic environment. To remove the interference of moving objects, we introduce a moving object segmentation (MOS) module. Besides, we design a scene graph transformer-attention module to generate a more discriminative and representative global scene graph descriptor, which significantly improves the performance of place recognition. In addition, we integrate our place recognition method for loop closure with an existing LiDAR-based odometry, boosting its localization accuracy. We evaluate our method on the KITTI and Oxford RobotCar datasets. Many experimental results show that our method can effectively accomplish place recognition, and its accuracy and robustness improve by at least 3% when compared with existing state-of-the-art methods, such as DiSCO. To illustrate the generalization capabilities of our method, we evaluate it on the KITTI-360 and NCLT datasets while using only KITTI for training. The experiments show that our scene graph descriptor can achieve accurate loop-closure and global localization in never-seen environments.
引用
收藏
页码:14 / 26
页数:13
相关论文
共 51 条
[1]   Automatic registration of optical aerial imagery to a LiDAR point cloud for generation of city models [J].
Abayowa, Bernard O. ;
Yilmaz, Alper ;
Hardie, Russell C. .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2015, 106 :68-81
[2]  
Arandjelovic R, 2018, IEEE T PATTERN ANAL, V40, P1437, DOI [10.1109/TPAMI.2017.2711011, 10.1109/CVPR.2016.572]
[3]   SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences [J].
Behley, Jens ;
Garbade, Martin ;
Milioto, Andres ;
Quenzel, Jan ;
Behnke, Sven ;
Stachniss, Cyrill ;
Gall, Juergen .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :9296-9306
[4]   University of Michigan North Campus long-term vision and lidar dataset [J].
Carlevaris-Bianco, Nicholas ;
Ushani, Arash K. ;
Eustice, Ryan M. .
INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2016, 35 (09) :1023-1035
[5]  
Chen X., 2021, arXiv
[6]  
Chen X., 2021, arXiv
[7]   Registration of large-scale terrestrial laser scanner point clouds: A review and benchmark [J].
Dong, Zhen ;
Liang, Fuxun ;
Yang, Bisheng ;
Xu, Yusheng ;
Zang, Yufu ;
Li, Jianping ;
Wang, Yuan ;
Dai, Wenxia ;
Fan, Hongchao ;
Hyyppa, Juha ;
Stilla, Uwe .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2020, 163 :327-342
[8]   A novel binary shape context for 3D local surface description [J].
Dong, Zhen ;
Yang, Bisheng ;
Liu, Yuan ;
Liang, Fuxun ;
Li, Bijun ;
Zang, Yufu .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2017, 130 :431-452
[9]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[10]  
Dube R., 2017, P IEEE INT C ROB AUT, P5266