RSMT: A Remote Sensing Image-to-Map Translation Model via Adversarial Deep Transfer Learning

被引:13
作者
Song, Jieqiong [1 ]
Li, Jun [1 ]
Chen, Hao [1 ]
Wu, Jiangjiang [1 ]
机构
[1] Natl Univ Def Technol, Coll Elect Sci & Technol, Changsha 410073, Peoples R China
基金
中国国家自然科学基金;
关键词
map translation; adversarial transfer learning; remote sensing image; attention mechanism;
D O I
10.3390/rs14040919
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Maps can help governments in infrastructure development and emergency rescue operations around the world. Using adversarial learning to generate maps from remote sensing images is an emerging field. As we now know, the urban construction styles of different cities are diverse. The current translation methods for remote sensing image-to-map tasks only work on the specific regions with similar styles and structures to the training set and perform poorly on previously unseen areas. We argue that this greatly limits their use. In this work, we intend to seek a remote sensing image-to-map translation model that approaches the challenge of generating maps for the remote sensing images of unseen areas. Our remote sensing image-to-map translation model (RSMT) achieves universal and general applicability to generate maps over multiple regions by combining adversarial deep transfer training schemes with novel attention-based network designs. Extracting the content and style latent features from remote sensing images and a series of maps, respectively, RSMT generalizes a pattern applied to the remote sensing images of new areas. Meanwhile, we introduce feature map loss and map consistency loss to reinforce generated maps' precision and geometry similarity. We critically analyze qualitative and quantitative results using widely adopted evaluation metrics through extensive validation and comparisons with previous remote sensing image-to-map approaches. The results of experiment indicate that RSMT can translate remote sensing images to maps better than several state-of-the-art methods.
引用
收藏
页数:20
相关论文
共 45 条
[1]  
Ajakan H., 2014, ARXIV14124446
[2]   Latent Filter Scaling for Multimodal Unsupervised Image-to-Image Translation [J].
Alharbi, Yazeed ;
Smith, Neil ;
Wonka, Peter .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :1458-1466
[3]  
Almahairi A, 2018, PR MACH LEARN RES, V80
[4]  
[Anonymous], 2015, P EMNLP, DOI DOI 10.18653/V1/D15-1166
[5]  
[Anonymous], 2018, P 2018 IEEE INT C AC
[6]  
Bahuleyan Hareesh, 2018, INT C COMP LING COLI, P1672
[7]   Attention to Scale: Scale-aware Semantic Image Segmentation [J].
Chen, Liang-Chieh ;
Yang, Yi ;
Wang, Jiang ;
Xu, Wei ;
Yuille, Alan L. .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :3640-3649
[8]  
Cheng Y, 2019, Joint Training for Neural Machine Translation, P11, DOI [10.1007/978-981-32-9748-7_2, DOI 10.1007/978-981-32-9748-7_2]
[9]   StarGAN v2: Diverse Image Synthesis for Multiple Domains [J].
Choi, Yunjey ;
Uh, Youngjung ;
Yoo, Jaejun ;
Ha, Jung-Woo .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :8185-8194
[10]   StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation [J].
Choi, Yunjey ;
Choi, Minje ;
Kim, Munyoung ;
Ha, Jung-Woo ;
Kim, Sunghun ;
Choo, Jaegul .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8789-8797