Remote Sensing Image Augmentation Based on Text Description for Waterside Change Detection

被引:18
作者
Chen, Chen [1 ]
Ma, Hongxiang [1 ]
Yao, Guorun [1 ]
Lv, Ning [2 ]
Yang, Hua [3 ]
Li, Cong [4 ]
Wan, Shaohua [5 ]
机构
[1] Xidian Univ, State Key Lab Integrated Serv Networks, Xian 710071, Peoples R China
[2] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[3] Northwest Univ, Sch Econ & Management, Xian 710127, Peoples R China
[4] State Grid JiLin Prov Elect Power Co Ltd Informat, Changchun 130000, Peoples R China
[5] Zhongnan Univ Econ & Law, Sch Informat & Safety Engn, Wuhan 430073, Peoples R China
基金
中国国家自然科学基金;
关键词
data augmentation; deeply monitoring; GAN; remote sensing image; text description;
D O I
10.3390/rs13101894
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Since remote sensing images are difficult to obtain and need to go through a complicated administrative procedure for use in China, it cannot meet the requirement of huge training samples for Waterside Change Detection based on deep learning. Recently, data augmentation has become an effective method to address the issue of an absence of training samples. Therefore, an improved Generative Adversarial Network (GAN), i.e., BTD-sGAN (Text-based Deeply-supervised GAN), is proposed to generate training samples for remote sensing images of Anhui Province, China. The principal structure of our model is based on Deeply-supervised GAN(D-sGAN), and D-sGAN is improved from the point of the diversity of the generated samples. First, the network takes Perlin Noise, image segmentation graph, and encoded text vector as input, in which the size of image segmentation graph is adjusted to 128 x 128 to facilitate fusion with the text vector. Then, to improve the diversity of the generated images, the text vector is used to modify the semantic loss of the downsampled text. Finally, to balance the time and quality of image generation, only a two-layer Unet++ structure is used to generate the image. Herein, "Inception Score", "Human Rank", and "Inference Time" are used to evaluate the performance of BTD-sGAN, StackGAN++, and GAN-INT-CLS. At the same time, to verify the diversity of the remote sensing images generated by BTD-sGAN, this paper compares the results when the generated images are sent to the remote sensing interpretation network and when the generated images are not added; the results show that the generated image can improve the precision of soil-moving detection by 5%, which proves the effectiveness of the proposed model.
引用
收藏
页数:18
相关论文
共 41 条
[1]  
Nguyen A, 2016, ADV NEUR IN, V29
[2]   Plug & Play Generative Networks: Conditional Iterative Generation of Images in Latent Space [J].
Anh Nguyen ;
Clune, Jeff ;
Bengio, Yoshua ;
Dosovitskiy, Alexey ;
Yosinski, Jason .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :3510-3520
[3]   Caching in Vehicular Named Data Networking: Architecture, Schemes and Future Directions [J].
Chen, Chen ;
Wang, Cong ;
Qiu, Tie ;
Atiquzzaman, Mohammed ;
Wu, Dapeng Oliver .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2020, 22 (04) :2378-2407
[4]   An Edge Traffic Flow Detection Scheme Based on Deep Learning in an Intelligent Transportation System [J].
Chen, Chen ;
Liu, Bin ;
Wan, Shaohua ;
Qiao, Peng ;
Pei, Qingqi .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (03) :1840-1852
[5]  
Denton E, 2015, ADV NEUR IN, V28
[6]   Stimulus-driven and concept-driven analysis for image caption generation [J].
Ding, Songtao ;
Qu, Shiru ;
Xi, Yuling ;
Wan, Shaohua .
NEUROCOMPUTING, 2020, 398 :520-530
[7]   Exploring Deep Learning for View-Based 3D Model Retrieval [J].
Gao, Zan ;
Li, Yinming ;
Wan, Shaohua .
ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2020, 16 (01)
[8]  
Gauthier JP, 2014, CONF P INDIUM PHOSPH
[9]   Change Detection within Remotely Sensed Satellite Image Time Series via Spectral Analysis [J].
Ghaderpour, Ebrahim ;
Vujadinovic, Tijana .
REMOTE SENSING, 2020, 12 (23) :1-27
[10]  
Goodfellow IJ, 2014, ADV NEUR IN, V27, P2672