LFSMIM: A Low-Frequency Spectral Masked Image Modeling Method for Hyperspectral Image Classification

被引:13
作者
Chen, Yuhan [1 ,2 ]
Yan, Qingyun [1 ]
机构
[1] Nanjing Univ Informat Sci & Technol, Sch Remote Sensing & Geomat Engn, Nanjing 210044, Peoples R China
[2] Harbin Engn Univ, Qingdao Innovat & Dev Base Ctr, Qingdao 266000, Peoples R China
基金
中国国家自然科学基金;
关键词
Image reconstruction; Transformers; Training; Discrete Fourier transforms; Principal component analysis; Feature extraction; Decoding; Hyperspectral image (HSI); masked image modeling (MIM); self-supervised learning; vision transformer (ViT);
D O I
10.1109/LGRS.2024.3360184
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Masked image modeling (MIM) has made significant advancements across various fields in recent years. Previous research in the hyperspectral (HS) domain often utilizes conventional Transformers to model spectral sequences, overlooking the impact of local details on HS image classification. Furthermore, training models using raw image features as reconstruction targets entail significant challenges. In this study, we specifically focus on the reconstruction targets and feature modeling capabilities of the Vision Transformer (ViT) to address the limitations of MIM methods in the HS domain. As a proposed solution, we introduce a novel and effective method called LFSMIM, which incorporates two key strategies: 1) filtering out high-frequency components from the reconstruction target to mitigate the network's sensitivity to noise and 2) enhancing the local and global modeling capabilities of the ViT to effectively capture weakened texture details and exploit global spectral features. LFSMIM demonstrated superior performance in overall accuracy (OA) compared to other methods on the Indian Pines (IP), Pavia University (PU), and Houston 2013 (HT) datasets, achieving accuracies of 95.522%, 98.820%, and 98.160% respectively. The code will be made available at https://github.com/yuweikong/LFSMIM.
引用
收藏
页码:1 / 5
页数:5
相关论文
共 15 条
[1]   Shallow-Guided Transformer for Semantic Segmentation of Hyperspectral Remote Sensing Imagery [J].
Chen, Yuhan ;
Liu, Pengyuan ;
Zhao, Jiechen ;
Huang, Kaijian ;
Yan, Qingyun .
REMOTE SENSING, 2023, 15 (13)
[2]   Deep Learning-Based Classification of Hyperspectral Data [J].
Chen, Yushi ;
Lin, Zhouhan ;
Zhao, Xing ;
Wang, Gang ;
Gu, Yanfeng .
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2014, 7 (06) :2094-2107
[3]  
Dosovitskiy A, 2021, Arxiv, DOI arXiv:2010.11929
[4]   Hyperspectral Image Classification Using a Hybrid 3D-2D Convolutional Neural Networks [J].
Ghaderizadeh, Saeed ;
Abbasi-Moghadam, Dariush ;
Sharifi, Alireza ;
Zhao, Na ;
Tariq, Aqil .
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 (14) :7570-7588
[5]   Masked Autoencoders Are Scalable Vision Learners [J].
He, Kaiming ;
Chen, Xinlei ;
Xie, Saining ;
Li, Yanghao ;
Dollar, Piotr ;
Girshick, Ross .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :15979-15988
[6]  
He MY, 2017, IEEE IMAGE PROC, P3904, DOI 10.1109/ICIP.2017.8297014
[7]   SpectralFormer: Rethinking Hyperspectral Image Classification With Transformers [J].
Hong, Danfeng ;
Han, Zhu ;
Yao, Jing ;
Gao, Lianru ;
Zhang, Bing ;
Plaza, Antonio ;
Chanussot, Jocelyn .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
[8]   Masked Auto-Encoding Spectral-Spatial Transformer for Hyperspectral Image Classification [J].
Ibanez, Damian ;
Fernandez-Beltran, Ruben ;
Pla, Filiberto ;
Yokoya, Naoto .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
[9]  
Lou M, 2023, Arxiv, DOI [arXiv:2310.19380, DOI 10.48550/ARXIV.2310.19380]
[10]   Hyperspectral Image Classification Using Group-Aware Hierarchical Transformer [J].
Mei, Shaohui ;
Song, Chao ;
Ma, Mingyang ;
Xu, Fulin .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60