MAP-Net: SAR and Optical Image Matching via Image-Based Convolutional Network With Attention Mechanism and Spatial Pyramid Aggregated Pooling

被引:47
作者
Cui, Song [1 ]
Ma, Ailong [1 ]
Zhang, Liangpei [1 ]
Xu, Miaozhong [1 ]
Zhong, Yanfei [1 ]
机构
[1] Wuhan Univ, State Key Lab Informat Engn Surveying Mapping & R, Wuhan 430079, Peoples R China
来源
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING | 2022年 / 60卷
基金
中国国家自然科学基金;
关键词
Feature extraction; Optical distortion; Image matching; Synthetic aperture radar; Optical sensors; Optical imaging; Nonlinear optics; Attention mechanism; convolutional neural network (CNN); deep learning; image matching; optical image; spatial pyramid pooling; synthetic aperture radar (SAR); DEEP; REGISTRATION;
D O I
10.1109/TGRS.2021.3066432
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
The complementarity of synthetic aperture radar (SAR) and optical images allows remote sensing observations to ``see'' unprecedented discoveries. Image matching plays a fundamental role in the fusion and application of SAR and optical images. However, both the geometric imaging pattern and the physical radiation mechanism of these two sensors are significantly different, so that the images show complex geometric distortion and nonlinear radiation differences. This phenomenon brings great challenges to image matching, which neither the handcrafted descriptors nor the deep learning-based methods have adequately addressed. In this article, a novel image-based matching method for SAR to optical images via an image-based convolutional network with spatial pyramid aggregated pooling (SPAP) and an attention mechanism is proposed, namely MAP-Net. The original image is embedded through the convolutional neural network to generate the feature map. Through the information extraction and abstraction of the original imagery, the embedded features containing the high-level semantic information are more robust to the geometric distortion and radiation variation among the different modal images, which is beneficial to the matching of cross-modal images. The adoption of the SPAP module makes the network more capable of integrating global and local contextual information. The attention block weights the dense features generated from the network to extract the key features that are invariant, distinguishable, repeatable, and suitable for the image matching task. In the experiments, five sets of multisource and multiresolution SAR and optical images with wide and varied ground coverage were used to evaluate the accuracy of MAP-Net, compared to both handcrafted and deep learning-based methods. The experimental results show that the MAP-Net method is superior to the current state-of-the-art image matching methods for SAR to optical images.
引用
收藏
页数:13
相关论文
共 53 条
[1]   Neural Codes for Image Retrieval [J].
Babenko, Artem ;
Slesarev, Anton ;
Chigorin, Alexandr ;
Lempitsky, Victor .
COMPUTER VISION - ECCV 2014, PT I, 2014, 8689 :584-599
[2]  
Bahdanau D, 2016, Arxiv, DOI [arXiv:1409.0473, DOI 10.48550/ARXIV.1409.0473]
[3]   HPatches: A benchmark and evaluation of handcrafted and learned local descriptors [J].
Balntas, Vassileios ;
Lenc, Karel ;
Vedaldi, Andrea ;
Mikolajczyk, Krystian .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3852-3861
[4]   MULTIDIMENSIONAL BINARY SEARCH TREES USED FOR ASSOCIATIVE SEARCHING [J].
BENTLEY, JL .
COMMUNICATIONS OF THE ACM, 1975, 18 (09) :509-517
[5]   A SURVEY OF IMAGE REGISTRATION TECHNIQUES [J].
BROWN, LG .
COMPUTING SURVEYS, 1992, 24 (04) :325-376
[6]   Vehicle Detection in Satellite Images by Hybrid Deep Convolutional Neural Networks [J].
Chen, Xueyun ;
Xiang, Shiming ;
Liu, Cheng-Lin ;
Pan, Chun-Hong .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2014, 11 (10) :1797-1801
[7]  
Curlander J. C., 1991, Synthetic Aperture Radar: Systems and Signal Processing, V11
[8]   SAR-SIFT: A SIFT-Like Algorithm for SAR Images [J].
Dellinger, Flora ;
Delon, Julie ;
Gousseau, Yann ;
Michel, Julien ;
Tupin, Florence .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2015, 53 (01) :453-466
[9]   SuperPoint: Self-Supervised Interest Point Detection and Description [J].
DeTone, Daniel ;
Malisiewicz, Tomasz ;
Rabinovich, Andrew .
PROCEEDINGS 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2018, :337-349
[10]  
Dugas C, 2001, ADV NEUR IN, V13, P472