Evaluating Task-Specific Augmentations in Self-Supervised Pre-Training for 3D Medical Image Analysis

被引:1
作者
Claessens, C. H. B. [1 ]
Hamm, J. J. M. [2 ]
Viviers, C. G. A. [1 ]
Nederend, J. [3 ]
Grunhagen, D. J. [2 ]
Tanis, P. J. [2 ]
de With, P. H. N. [1 ]
van der Sommen, F. [1 ]
机构
[1] Eindhoven Univ Technol, Eindhoven, Netherlands
[2] Erasmus MC, Rotterdam, Netherlands
[3] Catharina Hosp, Eindhoven, Netherlands
来源
MEDICAL IMAGING 2024: IMAGE PROCESSING | 2024年 / 12926卷
关键词
self-supervised learning; pre-training; medical imaging; three-dimensional; augmentations; self-distillation;
D O I
10.1117/12.3000850
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Self-supervised learning (SSL) has become a crucial approach for pre-training deep learning models in natural and medical image analysis. However, applying transformations designed for natural images to three-dimensional (3D) medical data poses challenges. This study explores the efficacy of specific augmentations in the context of self-supervised pre-training for volumetric medical images. A 3D non-contrastive framework is proposed for in-domain self-supervised pre-training on 3D gray-scale thorax CT data, incorporating four spatial and two intensity augmentations commonly used in 3D medical image analysis. The pre-trained models, adapted versions of ResNet-50 and Vision Transformer (ViT)-S, are evaluated on lung nodule classification and lung tumor segmentation tasks. The results indicate a significant impact of SSL, with a remarkable increase in AUC and DSC as compared to training from scratch. For classification, random scalings and random rotations play a fundamental role in achieving higher downstream performance, while intensity augmentations show limited contribution and may even degrade performance. For segmentation, random intensity histogram shifting enhances robustness, while other augmentations have marginal or negative impacts. These findings underscore the necessity of tailored data augmentations within SSL for medical imaging, emphasizing the importance of task-specific transformations for optimal model performance in complex 3D medical datasets.
引用
收藏
页数:8
相关论文
共 50 条
[31]   GUIDED CONTRASTIVE SELF-SUPERVISED PRE-TRAINING FOR AUTOMATIC SPEECH RECOGNITION [J].
Khare, Aparna ;
Wu, Minhua ;
Bhati, Saurabhchand ;
Droppo, Jasha ;
Maas, Roland .
2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, :174-181
[32]   Class incremental learning with self-supervised pre-training and prototype learning [J].
Liu, Wenzhuo ;
Wu, Xin-Jian ;
Zhu, Fei ;
Yu, Ming-Ming ;
Wang, Chuang ;
Liu, Cheng-Lin .
PATTERN RECOGNITION, 2025, 157
[33]   Masked Contrastive Representation Learning for Self-supervised Visual Pre-training [J].
Yao, Yuchong ;
Desai, Nandakishor ;
Palaniswami, Marimuthu .
2024 IEEE 11TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS, DSAA 2024, 2024, :9-18
[34]   DenseCL: A simple framework for self-supervised dense visual pre-training [J].
Wang, Xinlong ;
Zhang, Rufeng ;
Shen, Chunhua ;
Kong, Tao .
VISUAL INFORMATICS, 2023, 7 (01) :30-40
[35]   Self-supervised Pre-training with Acoustic Configurations for Replay Spoofing Detection [J].
Shim, Hye-jin ;
Heo, Hee-Soo ;
Jung, Jee-weon ;
Yu, Ha-Jin .
INTERSPEECH 2020, 2020, :1091-1095
[36]   Self-supervised multimodal reconstruction pre-training for retinal computer-aided diagnosis [J].
Hervella, Alvaro S. ;
Rouco, Jose ;
Novo, Jorge ;
Ortega, Marcos .
EXPERT SYSTEMS WITH APPLICATIONS, 2021, 185
[37]   A self-supervised pre-training scheme for multi-source heterogeneous remote sensing image land cover classification [J].
Xue Z. ;
Yu X. ;
Liu J. ;
Yang G. ;
Liu B. ;
Yu A. ;
Zhou J. ;
Jin S. .
Cehui Xuebao/Acta Geodaetica et Cartographica Sinica, 2024, 53 (03) :512-525
[38]   S3T: SELF-SUPERVISED PRE-TRAINING WITH SWIN TRANSFORMER FOR MUSIC CLASSIFICATION [J].
Zhao, Hang ;
Zhang, Chen ;
Zhu, Bilei ;
Ma, Zejun ;
Zhang, Kejun .
2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, :606-610
[39]   DeSD: Self-Supervised Learning with Deep Self-Distillation for 3D Medical Image Segmentation [J].
Ye, Yiwen ;
Zhang, Jianpeng ;
Chen, Ziyang ;
Xia, Yong .
MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT IV, 2022, 13434 :545-555
[40]   BRAVEN: IMPROVING SELF-SUPERVISED PRE-TRAINING FOR VISUAL AND AUDITORY SPEECH RECOGNITION [J].
Haliassos, Alexandros ;
Zinonos, Andreas ;
Mira, Rodrigo ;
Petridis, Stavros ;
Pantie, Maja .
2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2024), 2024, :11431-11435