Optimal Feature Transport for Cross-View Image Geo-Localization

被引:0
|
作者
Shi, Yujiao [1 ,2 ]
Yu, Xin [1 ,2 ]
Liu, Liu [1 ,2 ]
Zhang, Tong [1 ,3 ]
Li, Hongdong [1 ,2 ]
机构
[1] Australian Natl Univ, Canberra, ACT, Australia
[2] Australian Ctr Robot Vis, Canberra, ACT, Australia
[3] Motovis Australia Pty Ltd, Adelaide, SA, Australia
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper addresses the problem of cross-view image geo-localization, where the geographic location of a ground-level street-view query image is estimated by matching it against a large scale aerial map (e.g., a high-resolution satellite image). State-of-the-art deep-learning based methods tackle this problem as deep metric learning which aims to learn global feature representations of the scene seen by the two different views. Despite promising results are obtained by such deep metric learning methods, they, however, fail to exploit a crucial cue relevant for localization, namely, the spatial layout of local features. Moreover, little attention is paid to the obvious domain gap (between aerial view and ground view) in the context of cross-view localization. This paper proposes a novel Cross-View Feature Transport (CVFT) technique to explicitly establish cross-view domain transfer that facilitates feature alignment between ground and aerial images. Specifically, we implement the CVFT as network layers, which transports features from one domain to the other, leading to more meaningful feature similarity comparison. Our model is differentiable and can be learned end-to-end. Experiments on large-scale datasets have demonstrated that our method has remarkably boosted the state-of-the-art cross-view localization performance, e.g., on the CVUSA dataset, with significant improvements for top-1 recall from 40.79% to 61.43%, and for top-10 from 76.36% to 90.49%. We expect the key insight of the paper (i.e., explicitly handling domain difference via domain transport) will prove to be useful for other similar problems in computer vision as well.
引用
收藏
页码:11990 / 11997
页数:8
相关论文
共 50 条
  • [1] Cross-View Image Sequence Geo-localization
    Zhang, Xiaohan
    Sultani, Waqas
    Wshah, Safwan
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 2913 - 2922
  • [2] Feature Relation Guided Cross-View Image Based Geo-Localization
    Hou, Qingfeng
    Lu, Jun
    Guo, Haitao
    Liu, Xiangyun
    Gong, Zhihui
    Zhu, Kun
    Ping, Yifan
    REMOTE SENSING, 2023, 15 (20)
  • [3] Learning Robust Feature Representation for Cross-View Image Geo-Localization
    Gan, Wenjian
    Zhou, Yang
    Hu, Xiaofei
    Zhao, Luying
    Huang, Gaoshuang
    Hou, Mingbo
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2025, 22
  • [4] Cross-View Geo-Localization: A Survey
    Durgam, Abhilash
    Paheding, Sidike
    Dhiman, Vikas
    Devabhaktuni, Vijay
    IEEE ACCESS, 2024, 12 : 192028 - 192050
  • [5] Perceptual Feature Fusion Network for Cross-View Geo-Localization
    Wang, Jiayi
    Chen, Ziyang
    Yuan, Xiaochen
    Zhao, Genping
    Computer Engineering and Applications, 60 (03): : 255 - 262
  • [6] Spatial-Aware Feature Aggregation for Cross-View Image based Geo-Localization
    Shi, Yujiao
    Liu, Liu
    Yu, Xin
    Li, Hongdong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [7] Cross-View Image Matching for Geo-localization in Urban Environments
    Tian, Yicong
    Chen, Chen
    Shah, Mubarak
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 1998 - 2006
  • [8] Geographic Semantic Network for Cross-View Image Geo-Localization
    Zhu, Yingying
    Sun, Bin
    Lu, Xiufan
    Jia, Sen
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [9] GAMa: Cross-View Video Geo-Localization
    Vyas, Shruti
    Chen, Chen
    Shah, Mubarak
    COMPUTER VISION, ECCV 2022, PT XXXVII, 2022, 13697 : 440 - 456
  • [10] Cross-view geo-localization with evolving transformer
    Yang, Hongji
    Lu, Xiufan
    Zhu, Yingying
    arXiv, 2021,