Evaluation of deep learning algorithm for crop identification based on GF-6 time series images

被引:0
|
作者
Chen S. [1 ]
Liu J. [1 ]
机构
[1] Institute of Agricultural Resource and Regional Planning, Chinese Academy of Agricultural Sciences, Beijing
来源
Nongye Gongcheng Xuebao/Transactions of the Chinese Society of Agricultural Engineering | 2021年 / 37卷 / 15期
关键词
Crops; Deep learning; GF-6; Heilongjiang; Recognition; Remote sensing; Time series;
D O I
10.11975/j.issn.1002-6819.2021.15.020
中图分类号
学科分类号
摘要
Crop type mapping is one of the most important tools with medium and high spatial resolution satellite images in monitoring services of modern agriculture. Taking Heilongjiang Province of northeast China as a study area, this study aims to evaluate the state-of-the-art deep learning in crop type classification. A comparison was made on the Convolution Neural Network (CNN), Recurrent Neural Network (RNN), and Attention Mechanism (AM) for the application in crop type classification, while the traditional random forest (RF) model was also used as the control. Six models of deep learning were Temporal Convolutional Neural Network (TempCNN), Multi Scale 1D Residual Network (MSResNet), Omniscale Convolutional Neural Network (OmniscaleCNN), Long Short-Term Memory (LSTM), STAR Recurrent Neural Network (StarRNN), and Transformer. The specific procedure was as follows. First, GF-6 wide-field view image time series was acquired between April and November in the Lindian and Hailun study area, northeast of China, in order to extract the features of three types of crops at different growth stages. The resulting image time series used in the Lindian and the Hailun was composed of 41 and 48 GF-6 images, respectively. The preprocessing workflow included RPC correction, radiometric calibration, convert to top-of-atmospheric and surface reflectance using 6S atmospheric correction. The image interpolation and global min-max normalization were also applied to fill the empty pixel, further improving the convergence speed and stability of neural networks. The ground truth data was manually labelled using a field survey combined with GF-2 high-resolution image to generate datasets for train and evaluation. The datasets included six crops, such as rice, maize, soybean, water, urban and rest, covering 2 003 629 pixels in Lindian, 935 679 pixels in Hailun. Second, all models were trained and evaluated in Lindian, according to the differences between CNN, RNN, AM, and RF. All models achieved an overall accuracy of 93%-95%, and F1-score above 89%, 84%, and 97% for soybean, maize, and rice, respectively, where three major crops were from both study areas. Thirdly, the trained model in Lindian was transferred to that in Hailun, where the overall classification accuracy of each model declined between 7.2% to 41.0%, due to the differences of land cover classes and temporal composition of the data. Among CNNs, the accuracy of MSResNet barely changed to recognize three types of crops after transfer. Since OmniScaleCNN was automatically adjusted the size of the convolution filter, the accuracy of OmniScaleCNN after the transfer was better than that of TempCNN. A forget gate was utilized in the LSTM and StarRNN among RNNs, in order to avoid gradient explosion and disappearance in the classification, where the overall accuracy declined less than 10% after transfer. However, the accuracy of attention-based Transformer and RF dropped significantly. All models performed better on the distribution of three types of crops under the condition that the spatial location and temporal composition of data remain unchanged, in terms of visual analysis of classified images. Two CNN or RNN models were expected to accurately identify the general distribution of all land cover classes, under the varying spatial location and temporal composition. Furthermore, the run time of each deep learning was within 1 h, less than 6.2 times of random forest. Time consumption in the whole process was associated with the model training, as well as the image treatment for the Hailun study area covering an area of about 10 000 km². Correspondingly, deep learning presented a high-precision and large-scale crop mapping, in terms of classification accuracy and operating efficiency, particularly that the transfer learning performed better than before. © 2021, Editorial Department of the Transactions of the Chinese Society of Agricultural Engineering. All right reserved.
引用
收藏
页码:161 / 168
页数:7
相关论文
共 36 条
  • [1] Wang Pengxin, Xun Lan, Li Li, Et al., Extraction of planting areas of main crops based on sparse representation of time-series leaf area index, Journal of Remote Sensing, 24, 5, pp. 121-129, (2018)
  • [2] Zhao Hongwei, Chen Zhongxin, Liu Jia, Deep learning for crop classification of remote sensing data: Applications and challenges, Chinese Journal of Agricultural Resources and Regional Planning, 41, 2, pp. 35-49, (2020)
  • [3] Lecun Y, Bottou L., Gradient-based learning applied to document recognition, Proceedings of the IEEE, 86, 11, pp. 2278-2324, (1998)
  • [4] Hinton E, Williams J., Learning representations by back propagating errors, Nature, 323, 6088, pp. 533-536, (1986)
  • [5] Maggiori E, Tarabalka Y, Charpiat G, Et al., Convolutional neural networks for large-scale remote sensing image classification, IEEE Transactions on Geoscience and Remote Sensing, 55, 2, pp. 645-657, (2016)
  • [6] Postadjian T, Bris L, Sahbi H, Et al., Investigating the potential of deep neural networks for large-scale classification of very high resolution satellite images, Remote Sensing and Spatial Information Sciences, 31, 5, pp. 11-25, (2017)
  • [7] Zhao Fei, Zhang Wenkai, Yan Zhiyuan, Et al., Multi-feature map pyramid fusion deep network for semantic segmentation on remote sensing data, Journal of Electronics and Information Technology, 41, 10, pp. 44-50, (2019)
  • [8] Chen Yang, Fan Rongshuang, Wang Jingxue, Et al., Cloud detection of ZY-3 satellite remote sensing images based on deep learning, Acta Optica Sinica, 38, 1, pp. 32-42, (2018)
  • [9] Zhang Q, Yuan Q, Zeng C, Et al., Missing data reconstruction in remote sensing image with a unified spatial-temporal-spectral deep convolutional neural network, IEEE Transactions on Geoscience and Remote Sensing, 56, 8, pp. 4274-4288, (2018)
  • [10] Ozcelik F, Alganci U, Sertel E, Et al., Rethinking CNN-based pansharpening: Guided colorization of panchromatic images via GANS, IEEE Transactions on Geoscience and Remote Sensing, 59, 4, pp. 3486-3501, (2020)