Conditional Generative Adversarial Network-Based Training Sample Set Improvement Model for the Semantic Segmentation of High-Resolution Remote Sensing Images

被引:23
作者
Pan, Xin [1 ,2 ]
Zhao, Jian [1 ,2 ]
Xu, Jun [1 ,2 ]
机构
[1] Changchun Inst Technol, Sch Comp Technol & Engn, Changchun 130012, Peoples R China
[2] Jilin Prov Key Lab Changbai Hist Culture & VR Rec, Changchun 130012, Peoples R China
来源
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING | 2021年 / 59卷 / 09期
基金
中国国家自然科学基金;
关键词
Remote sensing; Image segmentation; Training; Semantics; Generative adversarial networks; Feature extraction; Generators; Condition generation; generative adversarial network (GAN); high-resolution remote sensing classification; sample generation; semantic segmentation; CONVOLUTIONAL NEURAL-NETWORK; DEEP LEARNING FRAMEWORK; CLASSIFICATION; EXTRACTION;
D O I
10.1109/TGRS.2020.3033816
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
To achieve high segmentation quality, deep semantic segmentation neural networks (DSSNNs) need to be trained on diverse direction, location, and neighboring category combinations for each pixel in the input-output image patches. Achieving this goal requires a large sample set. However, in many practical application scenarios, a very large training sample set is too expensive to achieve, or insufficient remote sensing image data are available. These limitations directly affect the quality of the results of DSSNNs. To address the above-mentioned problem, this article proposes a conditional generative adversarial network (CGAN)-based training sample set improvement model (CGAN-TSIM) for the semantic segmentation of high-resolution remote sensing images. In CGAN-TSIM, the generator model of the CGAN can generate a sample image when a ground-truth image is an input as a "condition." A condition generation mechanism is designed to create ground-truth images, and these ground-truth conditions are used to drive the CGAN to generate samples containing more diverse object combinations, directions, and locations. These generated images can be added to the original training sample set to improve their spatial information diversity. Rather than simply relying on passively finding samples that contain diverse spatial information, CGAN-TSIM extracts high-level spatial information from the original training images and actively generates new sample images. Experiments show that the samples generated by CGAN-TSIM can improve the quality of the sample set. Compared with other traditional methods, CGAN-TSIM enables better classification accuracy.
引用
收藏
页码:7854 / 7870
页数:17
相关论文
共 51 条
[1]  
[Anonymous], 2017, P IEEE C COMP VIS PA
[2]   Beyond RGB: Very high resolution urban remote sensing with multimodal deep networks [J].
Audebert, Nicolas ;
Le Saux, Bertrand ;
Lefevre, Sebastien .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2018, 140 :20-32
[3]   SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation [J].
Badrinarayanan, Vijay ;
Kendall, Alex ;
Cipolla, Roberto .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2017, 39 (12) :2481-2495
[4]   ResUNet-a: A deep learning framework for semantic segmentation of remotely sensed data [J].
Diakogiannis, Foivos, I ;
Waldner, Francois ;
Caccetta, Peter ;
Wu, Chen .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2020, 162 :94-114
[5]   Generative Adversarial Networks Based on Collaborative Learning and Attention Mechanism for Hyperspectral Image Classification [J].
Feng, Jie ;
Feng, Xueliang ;
Chen, Jiantong ;
Cao, Xianghai ;
Zhang, Xiangrong ;
Jiao, Licheng ;
Yu, Tao .
REMOTE SENSING, 2020, 12 (07)
[6]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672
[7]   Generative Adversarial Networks-Based Semi-Supervised Learning for Hyperspectral Image Classification [J].
He, Zhi ;
Liu, Han ;
Wang, Yiwen ;
Hu, Jie .
REMOTE SENSING, 2017, 9 (10)
[8]   Relation Network for Multilabel Aerial Image Classification [J].
Hua, Yuansheng ;
Mou, Lichao ;
Zhu, Xiao Xiang .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2020, 58 (07) :4558-4572
[9]   Automatic building extraction from high-resolution aerial images and LiDAR data using gated residual refinement network [J].
Huang, Jianfeng ;
Zhang, Xinchang ;
Xin, Qinchuan ;
Sun, Ying ;
Zhang, Pengcheng .
ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2019, 151 :91-105
[10]   Image-to-Image Translation with Conditional Adversarial Networks [J].
Isola, Phillip ;
Zhu, Jun-Yan ;
Zhou, Tinghui ;
Efros, Alexei A. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5967-5976