Hyperspectral Image Classification Using Multi-Scale Lightweight Transformer

被引:2
|
作者
Gu, Quan [1 ]
Luan, Hongkang [1 ]
Huang, Kaixuan [1 ]
Sun, Yubao [1 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Engn Res Ctr Digital Forens, Jiangsu Collaborat Innovat Ctr Atmospher Environm, Minist Educ, Nanjing 210044, Peoples R China
基金
中国国家自然科学基金;
关键词
hyperspectral image classification; multi-scale spectral attention; Transformer; long-range spectral dependence; SPARSE REPRESENTATION;
D O I
10.3390/electronics13050949
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The distinctive feature of hyperspectral images (HSIs) is their large number of spectral bands, which allows us to identify categories of ground objects by capturing discrepancies in spectral information. Convolutional neural networks (CNN) with attention modules effectively improve the classification accuracy of HSI. However, CNNs are not successful in capturing long-range spectral-spatial dependence. In recent years, Vision Transformer (VIT) has received widespread attention due to its excellent performance in acquiring long-range features. However, it requires calculating the pairwise correlation between token embeddings and has the complexity of the square of the number of tokens, which leads to an increase in the computational complexity of the network. In order to cope with this issue, this paper proposes a multi-scale spectral-spatial attention network with frequency-domain lightweight Transformer (MSA-LWFormer) for HSI classification. This method synergistically integrates CNN, attention mechanisms, and Transformer into the spectral-spatial feature extraction module and frequency-domain fused classification module. Specifically, the spectral-spatial feature extraction module employs a multi-scale 2D-CNN with multi-scale spectral attention (MS-SA) to extract the shallow spectral-spatial features and capture the long-range spectral dependence. In addition, The frequency-domain fused classification module designs a frequency-domain lightweight Transformer that employs the Fast Fourier Transform (FFT) to convert features from the spatial domain to the frequency domain, effectively extracting global information and significantly reducing the time complexity of the network. Experiments on three classic hyperspectral datasets show that MSA-LWFormer has excellent performance.
引用
收藏
页数:21
相关论文
共 50 条
  • [41] CSiT: A Multiscale Vision Transformer for Hyperspectral Image Classification
    He, Wenxuan
    Huang, Weiliang
    Liao, Shuhong
    Xu, Zhen
    Yan, Jingwen
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2022, 15 : 9266 - 9277
  • [42] Dual attention transformer network for hyperspectral image classification
    Shu, Zhenqiu
    Wang, Yuyang
    Yu, Zhengtao
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 127
  • [43] Cognitively-Inspired Multi-Scale Spectral-Spatial Transformer for Hyperspectral Image Super-Resolution
    Xu, Qin
    Liu, Shiji
    Liu, Jinpei
    Luo, Bin
    COGNITIVE COMPUTATION, 2024, 16 (01) : 377 - 391
  • [44] A multi-scale enhanced large-kernel attention transformer network for lightweight image super-resolution
    Chang, Kairong
    Jun, Sun
    Biao, Yang
    Hu, Mingzhi
    Yang, Junlong
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (03)
  • [45] Grouped multi-scale vision transformer for medical image segmentation
    Zexuan Ji
    Zheng Chen
    Xiao Ma
    Scientific Reports, 15 (1)
  • [46] Deep Multi-Scale Transformer for Remote Sensing Image Restoration
    Li, Yanting
    2024 5TH INTERNATIONAL CONFERENCE ON GEOLOGY, MAPPING AND REMOTE SENSING, ICGMRS 2024, 2024, : 138 - 142
  • [47] Multi-scale feature learning via residual dynamic graph convolutional network for hyperspectral image classification
    Chen, Rong
    Vivone, Gemine
    Li, Guanghui
    Dai, Chenglong
    Chanussot, Jocelyn
    INTERNATIONAL JOURNAL OF REMOTE SENSING, 2024, 45 (03) : 863 - 888
  • [48] Efficient evolutionary multi-scale spectral-spatial attention fusion network for hyperspectral image classification
    Zhang, Mengxuan
    Lei, Zhikun
    Liu, Long
    Ma, Kun
    Shang, Ronghua
    Jiao, Licheng
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 262
  • [49] UC-former: A multi-scale image deraining network using enhanced transformer
    Zhou, Weina
    Ye, Linhui
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2024, 248
  • [50] Multi-scale Convolutional Attention Fuzzy Broad Network for Few-Shot Hyperspectral Image Classification
    Hu, Xiaopei
    Zhao, Guixin
    Yuan, Lu
    Dong, Xiangjun
    Dong, Aimei
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT II, 2024, 15017 : 46 - 60