A Multiscale Framework With Unsupervised Learning for Remote Sensing Image Registration

被引:130
作者
Ye, Yuanxin [1 ,2 ]
Tang, Tengfeng [1 ,2 ]
Zhu, Bai [1 ,2 ]
Yang, Chao [1 ,2 ]
Li, Bo [3 ]
Hao, Siyuan [4 ]
机构
[1] Southwest Jiaotong Univ, Fac Geosci & Environm Engn, Chengdu 610031, Peoples R China
[2] Southwest Jiaotong Univ, State Prov Joint Engn Lab Spatial Informat Techno, Chengdu 611756, Peoples R China
[3] Northwestern Polytech Univ, Sch Elect & Informat, Xian 710129, Peoples R China
[4] Qingdao Univ Technol, Sch Informat & Control Engn, Qingdao 266033, Peoples R China
来源
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING | 2022年 / 60卷
基金
中国国家自然科学基金;
关键词
Feature extraction; Optical distortion; Optical imaging; Optical sensors; Distortion; Unsupervised learning; Image registration; multimodal images; multiscale framework; unsupervised learning; DEEP; NETWORK; ARCHITECTURE; AFFINE;
D O I
10.1109/TGRS.2022.3167644
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Registration for multisensor or multimodal image pairs with a large degree of distortions is a fundamental task for many remote sensing applications. To achieve accurate and low-cost remote sensing image registration, we propose a multiscale framework with unsupervised learning, named MU-Net. Without costly ground truth labels, MU-Net directly learns the end-to-end mapping from the image pairs to their transformation parameters. MU-Net stacks several deep neural network (DNN) models on multiple scales to generate a coarse-to-fine registration pipeline, which prevents the backpropagation from falling into a local extremum and resists significant image distortions. We design a novel loss function paradigm based on structural similarity, which makes MU-Net suitable for various types of multimodal images. MU-Net is compared with traditional feature-based and area-based methods, as well as supervised and other unsupervised learning methods on the optical-optical, optical-infrared, optical-synthetic aperture radar (SAR), and optical-map datasets. Experimental results show that MU-Net achieves more comprehensive and accurate registration performance between these image pairs with geometric and radiometric distortions. We share the code implemented by Pytorch at https://github.com/yeyuanxin110/MU-Net.
引用
收藏
页数:15
相关论文
共 53 条
[1]   VoxelMorph: A Learning Framework for Deformable Medical Image Registration [J].
Balakrishnan, Guha ;
Zhao, Amy ;
Sabuncu, Mert R. ;
Guttag, John ;
Dalca, Adrian, V .
IEEE TRANSACTIONS ON MEDICAL IMAGING, 2019, 38 (08) :1788-1800
[2]   MAGSAC: Marginalizing Sample Consensus [J].
Barath, Daniel ;
Matas, Jiri ;
Noskova, Jana .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :10189-10197
[3]   Speeded-Up Robust Features (SURF) [J].
Bay, Herbert ;
Ess, Andreas ;
Tuytelaars, Tinne ;
Van Gool, Luc .
COMPUTER VISION AND IMAGE UNDERSTANDING, 2008, 110 (03) :346-359
[4]  
Cao Xiaohuan, 2017, Med Image Comput Comput Assist Interv, V10433, P300, DOI 10.1007/978-3-319-66182-7_35
[5]   Histograms of oriented gradients for human detection [J].
Dalal, N ;
Triggs, B .
2005 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOL 1, PROCEEDINGS, 2005, :886-893
[6]   Unsupervised Learning for Fast Probabilistic Diffeomorphic Registration [J].
Dalca, Adrian V. ;
Balakrishnan, Guha ;
Guttag, John ;
Sabuncu, Mert R. .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2018, PT I, 2018, 11070 :729-738
[7]   A deep learning framework for unsupervised affine and deformable image registration [J].
de Vos, Bob D. ;
Berendsen, Floris F. ;
Viergever, Max A. ;
Sokooti, Hessam ;
Staring, Marius ;
Isgum, Ivana .
MEDICAL IMAGE ANALYSIS, 2019, 52 :128-143
[8]  
DeTone D., 2016, ARXIV160603798
[9]   Remote Sensing Object Detection Based on Receptive Field Expansion Block [J].
Dong, Xiaohu ;
Fu, Ruigang ;
Gao, Yinghui ;
Qin, Yao ;
Ye, Yuanxin ;
Li, Biao .
IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
[10]  
Dubrofsky E., 2009, Master Thesis