Point Cloud Registration Based on Adaptively Fused Multimodal Features

被引:0
作者
Shou, Yejun [1 ,2 ]
Wei, Xiaoyao [1 ,2 ]
Wang, Haocheng [1 ,2 ]
Zheng, Qian [3 ,4 ]
Li, Shuai [1 ,2 ]
Xu, Zhijie [5 ]
Pan, Gang [3 ,4 ]
Cao, Yanlong [1 ,2 ]
机构
[1] Zhejiang Univ, State Key Lab Fluid Power & Mechatron Syst, Hangzhou 310027, Peoples R China
[2] Zhejiang Univ, Key Lab Adv Mfg Technol Zhejiang Prov, Hangzhou 310027, Peoples R China
[3] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou 310027, Peoples R China
[4] Zhejiang Univ, State Key Lab Brain Machine Intelligence, Hangzhou 310027, Peoples R China
[5] Xian Jiaotong Liverpool Univ, Sch Adv Technol, Suzhou 215123, Peoples R China
关键词
Point cloud compression; Feature extraction; Three-dimensional displays; Benchmark testing; Data mining; Transformers; Robustness; Laser radar; Geometry; Fuses; Point cloud registration; geometrically weak scene understanding; sensor fusion; RGB-D perception; HISTOGRAMS;
D O I
10.1109/LRA.2025.3580326
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
Point cloud registration is a fundamental task in 3D vision, which plays an important role in various fields but faces challenges in geometrically weak or repetitive scenes. Traditional geometric-based methods struggle in these cases, while recent multimodal approaches improve robustness in weak scenes but rely on precise point cloud-image alignment, which is difficult in real-world, low-alignment environments. To address the above challenge, we propose the Adaptively Fused Multimodal Network (AFMNet). AFMNet establishes point-to-pixel correspondences at the sparse superpoint level and assigns weights to geometric and texture features, creating more distinct feature descriptors and reducing the impact of misalignment. Additionally, we introduce an image-guided confidence estimation strategy that assigns higher confidence levels to points within the alignment region, prioritizing their selection for registration. To better evaluate the robustness of point cloud registration methods in geometrically weak scenes, we build new benchmarks, 3DWeakMacth and 3DLoWeakMatch, based on 3DMatch and 3DLoMatch. Reasonable multimodal fusion enables our method to achieve state-of-the-art performance on both indoor 3DMatch, 3DLoMatch, 3DWeakMacth, and 3DLoWeakMatch benchmarks, as well as the outdoor KITTI benchmark with low alignment.
引用
收藏
页码:8163 / 8170
页数:8
相关论文
共 39 条
[1]   BUFFER: Balancing Accuracy, Efficiency, and Generalizability in Point Cloud Registration [J].
Ao, Sheng ;
Hu, Qingyong ;
Wang, Hanyun ;
Xu, Kai ;
Guo, Yulan .
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, :1255-1264
[2]   D3Feat: Joint Learning of Dense Detection and Description of 3D Local Features [J].
Bai, Xuyang ;
Luo, Zixin ;
Zhou, Lei ;
Fu, Hongbo ;
Quan, Long ;
Tai, Chiew-Lan .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :6358-6366
[3]   Moving Object Segmentation in 3D LiDAR Data: A Learning-Based Approach Exploiting Sequential Data [J].
Chen, Xieyuanli ;
Li, Shijie ;
Mersch, Benedikt ;
Wiesmann, Louis ;
Gall, Jurgen ;
Behley, Jens ;
Stachniss, Cyrill .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2021, 6 (04) :6529-6536
[4]   Fully Convolutional Geometric Features [J].
Choy, Christopher ;
Park, Jaesik ;
Koltun, Vladlen .
2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, :8957-8965
[5]   Learning-Based Point Cloud Registration for 6D Object Pose Estimation in the Real World [J].
Dang, Zheng ;
Wang, Lizhou ;
Guo, Yu ;
Salzmann, Mathieu .
COMPUTER VISION - ECCV 2022, PT I, 2022, 13661 :19-37
[6]   SuperFusion: Multilevel LiDAR-Camera Fusion for Long-Range HD Map Generation [J].
Dong, Hao ;
Gu, Weihao ;
Zhang, Xianjing ;
Xu, Jintao ;
Ai, Rui ;
Lu, Huimin ;
Kannala, Juho ;
Chen, Xieyuanli .
2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, :9056-9062
[7]   Bootstrap Your Own Correspondences [J].
El Banani, Mohamed ;
Johnson, Justin .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :6413-6422
[8]   RANDOM SAMPLE CONSENSUS - A PARADIGM FOR MODEL-FITTING WITH APPLICATIONS TO IMAGE-ANALYSIS AND AUTOMATED CARTOGRAPHY [J].
FISCHLER, MA ;
BOLLES, RC .
COMMUNICATIONS OF THE ACM, 1981, 24 (06) :381-395
[9]  
Geiger A, 2012, PROC CVPR IEEE, P3354, DOI 10.1109/CVPR.2012.6248074
[10]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778