Joint classification of hyperspectral and LiDAR data based on inter-modality match learning

被引:0
作者
Hang R. [1 ]
Sun Y. [1 ]
Liu Q. [1 ]
机构
[1] School of Computer, Nanjing University of Information Science and Technology
[2] 2. Jiangsu Collaborative Innovation Center of Atmospheric Environment and Equipment Technology
基金
中国国家自然科学基金;
关键词
deep learning; hyperspectral image; joint classification; LiDAR data; match learning; remote sensing image;
D O I
10.11834/jrs.20232635
中图分类号
学科分类号
摘要
Several excellent models for joint classification of hyperspectral image and LiDAR data, which were designed on the basis of supervised learning methods such as convolutional neural networks, have been developed in recent years. Their classification performance depends largely on the quantity and quality of training samples. However, when the distribution of ground objects becomes increasingly complex and the resolutions of remote sensing images grow increasingly high, obtaining high-quality labels only with limited cost and manpower is difficult. Therefore, numerous scholars have made efforts to learn features directly from unlabeled samples. For instance, the theory of autoencoder was applied to multimodal joint classification, achieving satisfactory performance. Methods based on the reconstruction idea reduce the dependence on labeled information to a certain extent, but several problems that must be settled still exist. For example, these methods pay more attention to data reconstruction but fail to guarantee that the extracted features have sufficient discriminant capability, thus affecting the performance of joint classification. This paper proposes an effective model named Joint Classification of Hyperspectral and LiDAR Data Based on Inter-Modality Match Learning to address the aforementioned issue. Different from feature extraction models based on reconstruction idea, the proposed model tends to compare the matching relationship between samples from different modalities, thereby enhancing the discriminative capability of features. Specifically, this model comprises inter-modality matching learning and multimodal joint classification networks. The former is prone to identify matching of the input patch pairs of hyperspectral image and LiDAR data; therefore, reasonable construction of matching labels is essential. Thus, spatial positions of center pixels in cropped patches and KMeans clustering methods are employed. These constructed labels and patch pairs are combined to train the network. Notably, this process does not use manual labeled information and can directly extract features from abundant unlabeled samples. Furthermore, in the joint classification stage, the structure and trained parameters of matching learning network are transferred, and a small number of manually labeled training samples are then used to finetune the model parameters. Extensive experiments were conducted on two widely used datasets, namely Houston and MUUFL, to verify the effectiveness of the proposed model. These experiments include comparison experiments with several state-of-the-art models, hyperparameter analysis experiments, and ablation studies. Take the first experiment as an example. Compared with other models, such as CNN, EMFNet, AE_H, AE_HL, CAE_H, CAE_HL, IP-CNN, and PToP CNN, the proposed model can achieve higher performance on both datasets with OAs of 88.39% and 81.46%, respectively. Overall, the proposed model reduces the dependence on manually labeled data and improves the joint classification accuracy in the case of limited training samples. A superior model structure and additional testing datasets will be explored in the future to make further improvements. © 2024 Science Press. All rights reserved.
引用
收藏
页码:154 / 167
页数:13
相关论文
共 32 条
[1]  
Abbasi A N, He M Y., Convolutional neural network with PCA and batch normalization for hyperspectral image classification, 2019 IEEE International Geoscience and Remote Sensing Symposium, pp. 959-962, (2019)
[2]  
Cao Q, Ma A L, Zhong Y F, Zhao J, Zhao B, Zhang L P., Urban classification by multi-feature fusion of hyperspectral image and LiDAR data, Journal of Remote Sensing, 23, 5, pp. 892-903, (2019)
[3]  
Chen Y S, Jiang H L, Li C Y, Jia X P, Ghamisi P., Deep feature extraction and classification of hyperspectral images based on convolutional neural networks, IEEE Transactions on Geoscience and Remote Sensing, 54, 10, pp. 6232-6251, (2016)
[4]  
Chen Y S, Li C Y, Ghamisi P, Jia X P, Gu Y F., Deep fusion of remote sensing data for accurate classification, IEEE Geoscience and Remote Sensing Letters, 14, 8, pp. 1253-1257, (2017)
[5]  
Chen Y S, Lin Z H, Zhao X, Wang G, Gu Y F., Deep learning-based classification of hyperspectral data, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 7, 6, pp. 2094-2107, (2014)
[6]  
Dalponte M, Bruzzone L, Gianelle D., Fusion of hyperspectral and LIDAR remote sensing data for classification of complex forest areas, IEEE Transactions on Geoscience and Remote Sensing, 46, 5, pp. 1416-1427, (2008)
[7]  
Debes C, Merentitis A, Heremans R, Hahn J, Frangiadakis N, Van Kasteren T, Liao W Z, Bellens R, Pizurica A, Gautama S, Philips W, Prasad S, Du Q, Pacifici F., Hyperspectral and LiDAR data fusion: outcome of the 2013 GRSS data fusion contest, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 7, 6, pp. 2405-2418, (2014)
[8]  
Gader P, Zare A, Close R, Aitken J, Tuell G., MUUFL Gulfport Hyperspectral and Lidar Airborne Data Set, (2013)
[9]  
Ge C R, Du Q, Sun W W, Wang K Y, Li J J, Li Y S., Deep residual network-based fusion framework for hyperspectral and LiDAR data, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 14, pp. 2458-2472, (2021)
[10]  
Hang R L, Li Z, Ghamisi P, Hong D F, Xia G Y, Liu Q S., Classification of hyperspectral and LiDAR Data using coupled CNNs, IEEE Transactions on Geoscience and Remote Sensing, 58, 7, pp. 4939-4950, (2020)