Cross-Site Visual Localization of Zhurong Mars Rover Based on Self-Supervised Keypoint Extraction and Robust Matching

被引:0
作者
Kou, Yuke [1 ]
Wan, Wenhui [1 ,2 ]
Di, Kaichang [1 ,2 ]
Liu, Zhaoqin [1 ,2 ]
Peng, Man [1 ,2 ]
Wang, Yexin [1 ,2 ]
Xie, Bin [1 ]
Wang, Biao [1 ]
Zhao, Chenxu [1 ]
机构
[1] Chinese Acad Sci, Aerosp Informat Res Inst, Key Lab Remote Sensing & Digital Earth, Beijing 100101, Peoples R China
[2] Univ Chinese Acad Sci, Coll Resources & Environm, Beijing 100049, Peoples R China
来源
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING | 2025年 / 63卷
关键词
Location awareness; Mars; Feature extraction; Space vehicles; Visualization; Accuracy; Training; Image matching; Data mining; Robustness; Cross-site visual localization; deep learning; feature matching; self-supervised training; Zhurong rover; ALGORITHM;
D O I
10.1109/TGRS.2025.3541152
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
High-precision localization of the Mars rovers is fundamental for path planning and safe navigation toward exploration targets during Mars missions. In cross-site visual localization, image matching is the key step to obtain corresponding points connecting images from different sites. The cross-site visual localization method based on Affine SIFT (ASIFT) is used in Tianwen-1 mission but is constrained in regions of Mars with poor texture and large viewpoint invariance. In this article, we propose a cross-site visual localization methodology of Mars rover based on self-supervised keypoint extraction and robust matching. The self-supervised keypoint extraction network, which is called MRSS-Net, uses multiscale deformable structures (MSDSs) during the feature encoding stage to enhance the network's ability of extracting invariant features in regions with large viewpoint variations and improve the rate of identical points for cross-site images with poor texture. In addition, we develop self-attention descriptor enhancement mechanism (SADEM) to distinguish local features in repetitive patterns. The robust matching, which is called adaptive 2-D-3-D matching, uses GNC dead-reckoning (3-D priori information) to construct the initial coarse matching domain and homography matrix (2-D information) to construct a progressively shrinking refined matching domain. We compared our method against ASIFT based cross-site visual localization model and advanced deep learning algorithms and evaluate the performance using NaTeCam images collected during the traversal of four long-distance traversals (a total of 44 Martian sol sites) by Zhurong rover. The experimental results show that our framework reduces the localization error by 12.5% and improves localization robustness by 50.8%, compared with ASIFT-based cross-site visual localization method used in Zhurong rover. In addition, our method outperforms state-of-the-art deep learning techniques and ensures the current accuracy of cross-site visual localization for Mars rover, while significantly increasing the level of automation.
引用
收藏
页数:20
相关论文
共 68 条
  • [1] KAZE Features
    Alcantarilla, Pablo Fernandez
    Bartoli, Adrien
    Davison, Andrew J.
    [J]. COMPUTER VISION - ECCV 2012, PT VI, 2012, 7577 : 214 - 227
  • [2] MAGSAC: Marginalizing Sample Consensus
    Barath, Daniel
    Matas, Jiri
    Noskova, Jana
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 10189 - 10197
  • [3] Graph-Cut RANSAC
    Barath, Daniel
    Matas, Jiri
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 6733 - 6741
  • [4] GMS: Grid-Based Motion Statistics for Fast, Ultra-robust Feature Correspondence
    Bian, Jia-Wang
    Lin, Wen-Yan
    Liu, Yun
    Zhang, Le
    Yeung, Sai-Kit
    Cheng, Ming-Ming
    Reid, Ian
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (06) : 1580 - 1593
  • [5] Biesiadecki JJ, 2006, AEROSP CONF PROC, P51
  • [6] An Approach to Science and Risk-Aware Planetary Rover Exploration
    Candela, Alberto
    Wettergreen, David
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2022, 7 (04): : 9691 - 9698
  • [7] Cavalli L, 2020, Arxiv, DOI [arXiv:2006.04250, DOI 10.48550/ARXIV.2006.04250]
  • [8] ASpanFormer: Detector-Free Image Matching with Adaptive Span Transformer
    Chen, Hongkai
    Luo, Zixin
    Zhou, Lei
    Tian, Yurun
    Zhen, Mingmin
    Fang, Tian
    McKinnon, David
    Tsin, Yanghai
    Quan, Long
    [J]. COMPUTER VISION - ECCV 2022, PT XXXII, 2022, 13692 : 20 - 36
  • [9] Matching with PROSAC - Progressive Sample Consensus
    Chum, O
    Matas, J
    [J]. 2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, : 220 - 226
  • [10] SuperPoint: Self-Supervised Interest Point Detection and Description
    DeTone, Daniel
    Malisiewicz, Tomasz
    Rabinovich, Andrew
    [J]. PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, : 337 - 349