Multimodal image matching: A scale-invariant algorithm and an open dataset

被引:19
作者
Li, Jiayuan [1 ]
Hu, Qingwu [1 ]
Zhang, Yongjun [1 ]
机构
[1] Wuhan Univ, Sch Remote Sensing & Informat Engn, Wuhan 430079, Peoples R China
关键词
Image matching; Feature descriptor; Dataset; SAR-optical; Multimodal images; AUTOMATIC REGISTRATION; FRAMEWORK; SAR; FUSION;
D O I
10.1016/j.isprsjprs.2023.08.010
中图分类号
P9 [自然地理学];
学科分类号
0705 ; 070501 ;
摘要
Multimodal image matching is a core basis for information fusion, change detection, and image-based navigation. However, multimodal images may simultaneously suffer from severe nonlinear radiation distortion (NRD) and complex geometric differences, which pose great challenges to existing methods. Although deep learning-based methods had shown potential in image matching, they mainly focus on same-source images or single types of multimodal images such as optical-synthetic aperture radar (SAR). One of the main obstacles is the lack of public data for different types of multimodal images. In this paper, we make two major contributions to the community of multimodal image matching: First, we collect six typical types of images, including optical-optical, optical-infrared, optical-SAR, optical-depth, optical-map, and nighttime, to construct a multimodal image dataset with a total of 1200 pairs. This dataset has good diversity in image categories, feature classes, resolutions, geometric variations, etc. Second, we propose a scale and rotation invariant feature transform (SRIF) method, which achieves good matching performance without relying on data characteristics. This is one of the advantages of our SRIF over deep learning methods. SRIF obtains the scales of FAST keypoints by projecting them into a simple pyramid scale space, which is based on the study that methods with/without scale space have similar performance under small scale change factors. This strategy largely reduces the complexity compared to traditional Gaussian scale space. SRIF also proposes a local intensity binary transform (LIBT) for SIFT-like feature description, which can largely enhance the structure information inside multimodal images. Extensive experiments on these 1200 image pairs show that our SRIF outperforms current state-of-the-arts by a large margin, including RIFT, CoFSM, LNIFT, and MS-HLMO. Both the created dataset and the code of SRIF will be publicly available in https://github.com/LJY-RS/SRIF.
引用
收藏
页码:77 / 88
页数:12
相关论文
共 76 条
  • [1] Aguilera CA, 2015, IEEE IMAGE PROC, P178, DOI 10.1109/ICIP.2015.7350783
  • [2] A NEW PRODUCTIVE FRAMEWORK FOR POINT-BASED MATCHING OF OBLIQUE AIRCRAFT AND UAV-BASED IMAGES
    Bas, Sila
    Ok, Ali Ozgun
    [J]. PHOTOGRAMMETRIC RECORD, 2021, 36 (175) : 252 - 284
  • [3] Speeded-Up Robust Features (SURF)
    Bay, Herbert
    Ess, Andreas
    Tuytelaars, Tinne
    Van Gool, Luc
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2008, 110 (03) : 346 - 359
  • [4] Brown M, 2011, PROC CVPR IEEE, P177, DOI 10.1109/CVPR.2011.5995637
  • [5] A Partial Intensity Invariant Feature Descriptor for Multimodal Retinal Image Registration
    Chen, Jian
    Tian, Jie
    Lee, Noah
    Zheng, Jian
    Smith, R. Theodore
    Laine, Andrew F.
    [J]. IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING, 2010, 57 (07) : 1707 - 1718
  • [6] Chenzhong gao, 2021, Journal of Beijing Institute of Technology, P113, DOI 10.15918/j.jbit1004-0579.2021.016
  • [7] Cho J., 2021, arXiv
  • [8] Modality-Free Feature Detector and Descriptor for Multimodal Remote Sensing Image Registration
    Cui, Song
    Xu, Miaozhong
    Ma, Ailong
    Zhong, Yanfei
    [J]. REMOTE SENSING, 2020, 12 (18)
  • [9] SuperPoint: Self-Supervised Interest Point Detection and Description
    DeTone, Daniel
    Malisiewicz, Tomasz
    Rabinovich, Andrew
    [J]. PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, : 337 - 349
  • [10] Infrared and Visible Image Registration Based on Scale-Invariant PIIFD Feature and Locality Preserving Matching
    Du, Qinglei
    Fan, Aoxiang
    Ma, Yong
    Fan, Fan
    Huang, Jun
    Mei, Xiaoguang
    [J]. IEEE ACCESS, 2018, 6 : 64107 - 64121