Spatial-Temporal Invariant Contrastive Learning for Remote Sensing Scene Classification

被引:14
作者
Huang, Haozhe [1 ]
Mou, Zhongfeng [1 ]
Li, Yunying [1 ]
Li, Qiujun [1 ]
Chen, Jie [1 ]
Li, Haifeng [1 ]
机构
[1] Cent South Univ, Sch Geosci & Info Phys, Changsha 410083, Peoples R China
基金
中国国家自然科学基金;
关键词
Task analysis; Remote sensing; Image analysis; Image color analysis; Training; Supervised learning; Histograms; Deep learning; remote sensing; scene classification; self-supervised learning; BENCHMARK;
D O I
10.1109/LGRS.2022.3173419
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Self-supervised learning achieves close to supervised learning results on remote sensing image (RSI) scene classification. This is due to the current popular self-supervised learning methods that learn representations by applying different augmentations to images and completing the instance discrimination task which enables convolutional neural networks (CNNs) to learn representations invariant to augmentation. However, RSIs are spatial-temporal heterogeneous, which means that similar features may exhibit different characteristics in different spatial-temporal scenes. Therefore, the performance of CNNs that learn only representations invariant to augmentation still degrades for unseen spatial-temporal scenes due to the lack of spatial-temporal invariant representations. We propose a spatial-temporal invariant contrastive learning (STICL) framework to learn spatial-temporal invariant representations from unlabeled images containing a large number of spatial-temporal scenes. We use optimal transport to transfer an arbitrary unlabeled RSI into multiple other spatial-temporal scenes and then use STICL to make CNNs produce similar representations for the views of the same RSI in different spatial-temporal scenes. We analyze the performance of our proposed STICL on four commonly used RSI scene classification datasets, and the results show that our method achieves better performance on RSIs in unseen spatial-temporal scenes compared to popular self-supervised learning methods. Based on our findings, it can be inferred that spatial-temporal invariance is an indispensable property for a remote sensing model that can be applied to a wider range of remote sensing tasks, which also inspires the study of more general remote sensing models. The source code is available at https://github.com/GeoX-Lab/G-RSIM/tree/main/T-SC-STICL.
引用
收藏
页数:5
相关论文
共 24 条
[1]  
Ayush K., 2021, PROC IEEECVF INT C C, P10
[2]   A data-driven adversarial examples recognition framework via adversarial feature genomes [J].
Chen, Li ;
Li, Qi ;
Chen, Weiye ;
Wang, Zeyu ;
Li, Haifeng .
INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (09) :6438-6462
[3]  
Chen T, 2020, PR MACH LEARN RES, V119
[4]  
Chen Y., 2021, arXiv
[5]   Remote Sensing Image Scene Classification: Benchmark and State of the Art [J].
Cheng, Gong ;
Han, Junwei ;
Lu, Xiaoqiang .
PROCEEDINGS OF THE IEEE, 2017, 105 (10) :1865-1883
[6]   MKN: Metakernel Networks for Few Shot Remote Sensing Scene Classification [J].
Cui, Zhenqi ;
Yang, Wang ;
Chen, Li ;
Li, Haifeng .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
[7]   Momentum Contrast for Unsupervised Visual Representation Learning [J].
He, Kaiming ;
Fan, Haoqi ;
Wu, Yuxin ;
Xie, Saining ;
Girshick, Ross .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :9726-9735
[8]   EuroSAT: A Novel Dataset and Deep Learning Benchmark for Land Use and Land Cover Classification [J].
Helber, Patrick ;
Bischke, Benjamin ;
Dengel, Andreas ;
Borth, Damian .
IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2019, 12 (07) :2217-2226
[9]  
Jing LL, 2022, INT J OCCUP SAF ERGO, V28, P842, DOI [10.1080/10803548.2020.1835234, 10.1109/TPAMI.2020.2991050]
[10]   Deep Unsupervised Embedding for Remotely Sensed Images Based on Spatially Augmented Momentum Contrast [J].
Kang, Jian ;
Fernandez-Beltran, Ruben ;
Duan, Puhong ;
Liu, Sicong ;
Plaza, Antonio J. .
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2021, 59 (03) :2598-2610