GMIM: Self-supervised pre-training for 3D medical image segmentation with adaptive and hierarchical masked image modeling

被引:1
|
作者
Qi L. [1 ]
Jiang Z. [1 ,2 ]
Shi W. [1 ,2 ]
Qu F. [1 ]
Feng G. [1 ]
机构
[1] Department of Computer Science and Technology, Changchun University of Science and Technology, Jilin, Changchun
[2] Zhongshan Institute of Changchun University of Science and Technology, Guangzhou, Zhongshan
关键词
Brain tumor segmentation; Masked image modeling; Self-supervised learning;
D O I
10.1016/j.compbiomed.2024.108547
中图分类号
学科分类号
摘要
Self-supervised pre-training and fully supervised fine-tuning paradigms have received much attention to solve the data annotation problem in deep learning fields. Compared with traditional pre-training on large natural image datasets, medical self-supervised learning methods learn rich representations derived from unlabeled data itself thus avoiding the distribution shift between different image domains. However, nowadays state-of-the-art medical pre-training methods were specifically designed for downstream tasks making them less flexible and difficult to apply to new tasks. In this paper, we propose grid mask image modeling, a flexible and general self-supervised method to pre-train medical vision transformers for 3D medical image segmentation. Our goal is to guide networks to learn the correlations between organs and tissues by reconstructing original images based on partial observations. The relationships are consistent within the human body and invariant to disease type or imaging modality. To achieve this, we design a Siamese framework consisting of an online branch and a target branch. An adaptive and hierarchical masking strategy is employed in the online branch to (1) learn the boundaries or small contextual mutation regions within images; (2) to learn high-level semantic representations from deeper layers of the multiscale encoder. In addition, the target branch provides representations for contrastive learning to further reduce representation redundancy. We evaluate our method through segmentation performance on two public datasets. The experimental results demonstrate our method outperforms other self-supervised methods. Codes are available at https://github.com/mobiletomb/Gmim. © 2024 Elsevier Ltd
引用
收藏
相关论文
共 50 条
  • [1] Intra-modality masked image modeling: A self-supervised pre-training method for brain tumor segmentation
    Qi, Liangce
    Shi, Weili
    Miao, Yu
    Li, Yonghui
    Feng, Guanyuan
    Jiang, Zhengang
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 95
  • [2] Evaluating Task-Specific Augmentations in Self-Supervised Pre-Training for 3D Medical Image Analysis
    Claessens, C. H. B.
    Hamm, J. J. M.
    Viviers, C. G. A.
    Nederend, J.
    Grunhagen, D. J.
    Tanis, P. J.
    de With, P. H. N.
    van der Sommen, F.
    MEDICAL IMAGING 2024: IMAGE PROCESSING, 2024, 12926
  • [3] Self-supervised 3D Anatomy Segmentation Using Self-distilled Masked Image Transformer (SMIT)
    Jiang, Jue
    Tyagi, Neelam
    Tringale, Kathryn
    Crane, Christopher
    Veeraraghavan, Harini
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT IV, 2022, 13434 : 556 - 566
  • [4] Self-supervised Pre-training for Nuclei Segmentation
    Haq, Mohammad Minhazul
    Huang, Junzhou
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT II, 2022, 13432 : 303 - 313
  • [5] ReFs: A hybrid pre-training paradigm for 3D medical image segmentation
    Xie, Yutong
    Zhang, Jianpeng
    Liu, Lingqiao
    Wang, Hu
    Ye, Yiwen
    Verjans, Johan
    Xia, Yong
    MEDICAL IMAGE ANALYSIS, 2024, 91
  • [6] Masked Deformation Modeling for Volumetric Brain MRI Self-Supervised Pre-Training
    Lyu, Junyan
    Bartlett, Perry F.
    Nasrallah, Fatima A.
    Tang, Xiaoying
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2025, 44 (03) : 1596 - 1607
  • [7] Self-Supervised Underwater Image Generation for Underwater Domain Pre-Training
    Wu, Zhiheng
    Wu, Zhengxing
    Chen, Xingyu
    Lu, Yue
    Yu, Junzhi
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2024, 73 : 1 - 14
  • [8] A Unified Visual Information Preservation Framework for Self-supervised Pre-Training in Medical Image Analysis
    Zhou, Hong-Yu
    Lu, Chixiang
    Chen, Chaoqi
    Yang, Sibei
    Yu, Yizhou
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (07) : 8020 - 8035
  • [9] Masked Text Modeling: A Self-Supervised Pre-training Method for Scene Text Detection
    Wang, Keran
    Xie, Hongtao
    Wang, Yuxin
    Zhang, Dongming
    Qu, Yadong
    Gao, Zuan
    Zhang, Yongdong
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 2006 - 2015
  • [10] GO-MAE: Self-supervised pre-training via masked autoencoder for OCT image classification of gynecology
    Wang, Haoran
    Guo, Xinyu
    Song, Kaiwen
    Sun, Mingyang
    Shao, Yanbin
    Xue, Songfeng
    Zhang, Hongwei
    Zhang, Tianyu
    NEURAL NETWORKS, 2025, 181