SMFE-Net: a saliency multi-feature extraction framework for VHR remote sensing image classification

被引:0
作者
Junsong Chen
Jizheng Yi
Aibin Chen
Ke Yang
Ze Jin
机构
[1] Central South University of Forestry and Technology,College of Computer and Information Engineering
[2] Central South University of Forestry and Technology,Institute of Artificial Intelligence Application
[3] Hunan Key Laboratory of Intelligent Logistics Technology,Laboratory for Future Interdisciplinary Research of Science and Technology, Institute of Innovative Research
[4] Suzuki Lab,undefined
[5] Information and Artificial Intelligence Research International Hub Group,undefined
[6] Tokyo Institute of Technology,undefined
来源
Multimedia Tools and Applications | 2024年 / 83卷
关键词
Very-high resolution (VHR) remote sensing image; Saliency multi-feature extraction; Attention mechanism; Remote sensing scene classification;
D O I
暂无
中图分类号
学科分类号
摘要
Scene classification of very-high resolution (VHR) remote sensing images is a challenging research hotspot. It is difficult to extract salient features because of the characteristics of remote sensing images, such as large spatial range changes and complex scenes. In addition, the effective combination of high-level semantic information and low-level contour information is also a major difficulty at present. In order to solve these problems, we proposed a new end-to-end saliency multi-feature extraction network (SMFE-Net) based on VGG16 and long short-term memory (LSTM) to extract salient features and effectively integrate high-level features with low-level features. Firstly, we design an adaptive memory network (AMN) based on the rectangular combination of LSTMs to capture rich features of high and low levels. The AMN not only provides supplementary information but also focuses on key areas, thus discarding non-critical information. Secondly, in order to realize adaptive feature extraction, the sequential connection of channel attention (CA) and spatial attention (SA) is placed in the high-level feature extraction subnetwork, whose outputs are multiplied by the weights of the last feature map layer of VGG16. Finally, the outputs of AMN and the attention-weighted features are concatenated and inputted to the fully connected layer for the scene classification of the VHR remote sensing image. To verify the validity of the proposed SMFE-Net, the UC Merced (UCM) land-use dataset, the Aerial Image Dataset (AID), and the OPTIMAL-31 (OPTL) dataset are selected as the experimental materials. Experimental results have demonstrated that the proposed SMFE-Net is superior to several most advanced methods.
引用
收藏
页码:3831 / 3854
页数:23
相关论文
共 138 条
[21]  
Benediktsson JA(2019)Learning context using segment-level LSTM for neural sequence labeling IEEE/ACM Trans Audio Speech Lang Process 28 105-96
[22]  
Guo D(2019)Remote sensing scene classification by gated bidirectional network IEEE Trans Geosci Remote Sens 58 82-1167
[23]  
Xia Y(2018)Scene classification with recurrent attention of VHR Remote sensing images IEEE Trans Geosci Remote Sens 57 1155-7932
[24]  
Luo X(2021)Enhanced feature pyramid network with deep semantic embedding for remote sensing scene classification IEEE Trans Geosci Remote Sens 59 7918-3981
[25]  
Guo Y(2021)Hierarchical object detection for very high-resolution satellite images Appl Soft Comput 113 107885-11
[26]  
Ji J(2021)Sentiment classification using attention mechanism and bidirectional long short-term memory network Appl Soft Comput 112 107792-1440
[27]  
Shi D(2017)AID: a benchmark dataset for performance evaluation of aerial scene classification IEEE Trans Geosci Remote Sens 55 3965-1334
[28]  
Ye Q(2021)Dual-channel residual network for hyperspectral image classification with noisy labels IEEE Trans Geosci Remote Sens 60 1-undefined
[29]  
Xie H(2022)ConDinet++: full-scale fusion network based on conditional dilated convolution to extract roads from remote sensing images IEEE Geosci Remote Sens Lett 19 8015105-undefined
[30]  
He N(2022)ConvPatchTrans: A script identification network with global and local semantics deeply integrated Eng Appl Artif Intell 113 104916-undefined