Interpretive Self-Supervised Pre-training: Boosting Performance on Visual Medical Data

被引:1
作者
Manna, Siladittya [1 ]
Bhattacharya, Saumik [2 ]
Pal, Umapada [1 ]
机构
[1] Indian Stat Inst, Kolkata, W Bengal, India
[2] Indian Inst Technol Kharagpur, Kharagpur, W Bengal, India
来源
PROCEEDINGS OF THE TWELFTH INDIAN CONFERENCE ON COMPUTER VISION, GRAPHICS AND IMAGE PROCESSING, ICVGIP 2021 | 2021年
关键词
Self-Supervised Learning; Contrastive Learning; Cosine Similarity; Loss function; Lower bound; Pre-training; Medical Data; CONTEXT;
D O I
10.1145/3490035.3490273
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-supervised learning algorithms have become one of the best tools for unsupervised representation learning. Although self-supervised algorithms have achieved state-of-the-art performance for classification tasks in the case of natural image data, their application on medical data has been limited. In this work, we have proposed a novel loss function and derive it's asymptotic lower bound. We have also shown that self-supervised pre-training with the proposed loss function helps in surpassing the supervised baseline on the downstream task. We have also shown that the self-supervised pretraining helps a model in learning better representation in general to achieve better performance compared to supervised baselines. We have mathematically derived that the contrastive loss function asymptotically treats each sample as a separate class and works by maximizing the distance between any two samples and this helps to get better performance. Finally, through exhaustive experiments, we demonstrate that self-supervised pre-training helps to surpass the performance of fully supervised models on downstream tasks.
引用
收藏
页数:9
相关论文
共 41 条
[11]  
DeVries T, 2017, Arxiv, DOI arXiv:1708.04552
[12]   Unsupervised Visual Representation Learning by Context Prediction [J].
Doersch, Carl ;
Gupta, Abhinav ;
Efros, Alexei A. .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1422-1430
[13]  
El-Nouby A, 2019, Arxiv, DOI arXiv:1910.12770
[14]   Self-Supervised Video Representation Learning With Odd-One-Out Networks [J].
Fernando, Basura ;
Bilen, Hakan ;
Gavves, Efstratios ;
Gould, Stephen .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5729-5738
[15]  
Godbole S, 2004, LECT NOTES ARTIF INT, V3056, P22
[16]  
Grill Jean-Bastien, 2020, ADV NEURAL INFORM PR, P21271, DOI [DOI 10.48550/ARXIV.2006.07733, 10.48550/arXiv.2006.07733]
[17]   Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1026-1034
[18]   Momentum Contrast for Unsupervised Visual Representation Learning [J].
He, Kaiming ;
Fan, Haoqi ;
Wu, Yuxin ;
Xie, Saining ;
Girshick, Ross .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :9726-9735
[19]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[20]  
Henaff OJ, 2020, PR MACH LEARN RES, V119