Self-Supervised Encoders Are Better Transfer Learners in Remote Sensing Applications

被引:4
作者
Calhoun, Zachary D. [1 ]
Lahrichi, Saad [2 ]
Ren, Simiao [3 ]
Malof, Jordan M. [4 ]
Bradbury, Kyle [3 ,5 ]
机构
[1] Duke Univ, Dept Civil & Environm Engn, 121 Hudson Hall, Durham, NC 27708 USA
[2] Duke Kunshan Univ, Div Nat & Appl Sci, 8 Duke Ave, Kunshan 215316, Peoples R China
[3] Duke Univ, Dept Elect & Comp Engn, Durham, NC 27708 USA
[4] Univ Montana, Dept Comp Sci, Missoula, MT 59812 USA
[5] Duke Univ, Nicholas Inst Energy Environm & Sustainabil, Durham, NC 27708 USA
关键词
machine learning; self-supervision; computer vision; transfer learning; SwAV; semantic segmentation; satellite imagery; domain adaptation;
D O I
10.3390/rs14215500
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Transfer learning has been shown to be an effective method for achieving high-performance models when applying deep learning to remote sensing data. Recent research has demonstrated that representations learned through self-supervision transfer better than representations learned on supervised classification tasks. However, little research has focused explicitly on applying self-supervised encoders to remote sensing tasks. Using three diverse remote sensing datasets, we compared the performance of encoders pre-trained through both supervision and self-supervision on ImageNet, then fine-tuned on a final remote sensing task. Furthermore, we explored whether performance benefited from further pre-training on remote sensing data. Our experiments used SwAV due to its comparably lower computational requirements, as this method would prove most easily replicable by practitioners. We show that an encoder pre-trained on ImageNet using self-supervision transfers better than one pre-trained using supervision on three diverse remote sensing applications. Moreover, self-supervision on the target data alone as a pre-training step seldom boosts performance beyond this transferred encoder. We attribute this inefficacy to the lower diversity and size of remote sensing datasets, compared to ImageNet. In conclusion, we recommend that researchers use self-supervised representations for transfer learning on remote sensing data and that future research should focus on ways to increase performance further using self-supervision.
引用
收藏
页数:14
相关论文
共 34 条
[1]  
[Anonymous], 2018, INT C LEARNING REPRE
[2]   Geography-Aware Self-Supervised Learning [J].
Ayush, Kumar ;
Uzkent, Burak ;
Meng, Chenlin ;
Tanmay, Kumar ;
Burke, Marshall ;
Lobell, David ;
Ermon, Stefano .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :10161-10170
[3]   Distributed solar photovoltaic array location and extent dataset for remote sensing object identification [J].
Bradbury, Kyle ;
Saboo, Raghav ;
Johnson, Timothy L. ;
Malof, Jordan M. ;
Devarajan, Arjun ;
Zhang, Wuming ;
Collins, Leslie M. ;
Newell, Richard G. .
SCIENTIFIC DATA, 2016, 3
[4]  
Caron M, 2020, ADV NEUR IN, V33
[5]   Emerging Properties in Self-Supervised Vision Transformers [J].
Caron, Mathilde ;
Touvron, Hugo ;
Misra, Ishan ;
Jegou, Herve ;
Mairal, Julien ;
Bojanowski, Piotr ;
Joulin, Armand .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :9630-9640
[6]  
Chen T., 2020, ADV NEURAL INF PROCE, V33, P22243
[7]  
Chen T, 2020, PR MACH LEARN RES, V119
[8]   Convolutional Neural Network for Remote-Sensing Scene Classification: Transfer Learning Analysis [J].
de Lima, Rafael Pires ;
Marfurt, Kurt .
REMOTE SENSING, 2020, 12 (01)
[9]   How Well Do Self-Supervised Models Transfer? [J].
Ericsson, Linus ;
Gouk, Henry ;
Hospedales, Timothy M. .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, :5410-5419
[10]  
Grill J.-B., 2020, P 34 INT C NEURAL IN, V33, P21271