SMFE-Net: a saliency multi-feature extraction framework for VHR remote sensing image classification

被引:0
作者
Junsong Chen
Jizheng Yi
Aibin Chen
Ke Yang
Ze Jin
机构
[1] Central South University of Forestry and Technology,College of Computer and Information Engineering
[2] Central South University of Forestry and Technology,Institute of Artificial Intelligence Application
[3] Hunan Key Laboratory of Intelligent Logistics Technology,Laboratory for Future Interdisciplinary Research of Science and Technology, Institute of Innovative Research
[4] Suzuki Lab,undefined
[5] Information and Artificial Intelligence Research International Hub Group,undefined
[6] Tokyo Institute of Technology,undefined
来源
Multimedia Tools and Applications | 2024年 / 83卷
关键词
Very-high resolution (VHR) remote sensing image; Saliency multi-feature extraction; Attention mechanism; Remote sensing scene classification;
D O I
暂无
中图分类号
学科分类号
摘要
Scene classification of very-high resolution (VHR) remote sensing images is a challenging research hotspot. It is difficult to extract salient features because of the characteristics of remote sensing images, such as large spatial range changes and complex scenes. In addition, the effective combination of high-level semantic information and low-level contour information is also a major difficulty at present. In order to solve these problems, we proposed a new end-to-end saliency multi-feature extraction network (SMFE-Net) based on VGG16 and long short-term memory (LSTM) to extract salient features and effectively integrate high-level features with low-level features. Firstly, we design an adaptive memory network (AMN) based on the rectangular combination of LSTMs to capture rich features of high and low levels. The AMN not only provides supplementary information but also focuses on key areas, thus discarding non-critical information. Secondly, in order to realize adaptive feature extraction, the sequential connection of channel attention (CA) and spatial attention (SA) is placed in the high-level feature extraction subnetwork, whose outputs are multiplied by the weights of the last feature map layer of VGG16. Finally, the outputs of AMN and the attention-weighted features are concatenated and inputted to the fully connected layer for the scene classification of the VHR remote sensing image. To verify the validity of the proposed SMFE-Net, the UC Merced (UCM) land-use dataset, the Aerial Image Dataset (AID), and the OPTIMAL-31 (OPTL) dataset are selected as the experimental materials. Experimental results have demonstrated that the proposed SMFE-Net is superior to several most advanced methods.
引用
收藏
页码:3831 / 3854
页数:23
相关论文
共 138 条
[11]  
Jin Z(2020)Deep discriminative representation learning with attention map for scene classification Remote Sens 12 1366-3702
[12]  
Cheng G(2020)Deepmultiple instance convolutional neural networks for learning robust scene representations IEEE Trans Geosci Remote Sens 58 3685-112654
[13]  
Yang C(2021)Generating and measuring similar sentences using long short-term memory and generative adversarial networks in IEEE Access 9 112637-648
[14]  
Yao X(2021)Sparse spatial attention network for semantic segmentation IEEE Int Conf Image Process (ICIP) 19 644-7121
[15]  
Guo L(2018)Scene classification based on multiscale convolutional neural network IEEE Trans Geosci Remote Sens 56 7109-160
[16]  
Han J(2017)Local binary features for texture classification: taxonomy and experimental study Pattern Recognit 62 135-146
[17]  
Fang L(2019)Bidirectional adaptive feature fusion for remote sensing scene classification Neurocomputing 328 135-7906
[18]  
He N(2019)A feature aggregation convolutional neuralnetwork for remote sensing scene classification IEEE Trans Geosci Remote Sens 57 7894-2195
[19]  
Li S(2017)Exploring models and data for remote sensing image caption generation IEEE Trans Geosci Remote Sens 56 2183-117909
[20]  
Ghamisi P(2021)Bloom’s Learning Outcomes’ automatic classification using LSTM and pretrained word embeddings IEEE Access 9 117887-115