MS-DINO: Masked Self-Supervised Distributed Learning Using Vision Transformer

被引:1
|
作者
Park, Sangjoon [1 ,2 ,3 ]
Lee, Ik Jae [4 ]
Kim, Jun Won [4 ]
Ye, Jong Chul [5 ]
机构
[1] Korea Adv Inst Sci & Technol, Dept Bio & Brain Engn, Daejeon 34141, South Korea
[2] Yonsei Univ, Coll Med, Dept Radiat Oncol, Seoul 03722, South Korea
[3] Yonsei Univ, Inst Innovat Digital Healthcare, Seoul 03722, South Korea
[4] Gangnam Severance Hosp, Dept Radiat Oncol, Seoul 06273, South Korea
[5] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch AI, Daejeon 34141, South Korea
基金
新加坡国家研究基金会;
关键词
Feature extraction; Task analysis; Biomedical imaging; Privacy; Transformers; Servers; Distance learning; Distributed learning; self-supervised learning; random permutation; vision transformer; privacy protection;
D O I
10.1109/JBHI.2024.3423797
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Despite promising advancements in deep learning in medical domains, challenges still remain owing to data scarcity, compounded by privacy concerns and data ownership disputes. Recent explorations of distributed-learning paradigms, particularly federated learning, have aimed to mitigate these challenges. However, these approaches are often encumbered by substantial communication and computational overhead, and potential vulnerabilities in privacy safeguards. Therefore, we propose a self-supervised masked sampling distillation technique called MS-DINO, tailored to the vision transformer architecture. This approach removes the need for incessant communication and strengthens privacy using a modified encryption mechanism inherent to the vision transformer while minimizing the computational burden on client-side devices. Rigorous evaluations across various tasks confirmed that our method outperforms existing self-supervised distributed learning strategies and fine-tuned baselines.
引用
收藏
页码:6180 / 6192
页数:13
相关论文
共 50 条
  • [31] Unsupervised Visual Anomaly Detection Using Self-Supervised Pre-Trained Transformer
    Kim, Jun-Hyung
    Kwon, Goo-Rak
    IEEE ACCESS, 2024, 12 : 127604 - 127613
  • [32] STFNet: Self-Supervised Transformer for Infrared and Visible Image Fusion
    Liu, Qiao
    Pi, Jiatian
    Gao, Peng
    Yuan, Di
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (02): : 1513 - 1526
  • [33] Self-supervised 3D Anatomy Segmentation Using Self-distilled Masked Image Transformer (SMIT)
    Jiang, Jue
    Tyagi, Neelam
    Tringale, Kathryn
    Crane, Christopher
    Veeraraghavan, Harini
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT IV, 2022, 13434 : 556 - 566
  • [34] Multi-Feature Vision Transformer via Self-Supervised Representation Learning for Improvement of COVID-19 Diagnosis
    Qi, Xiao
    Foran, David J.
    Nosher, John L.
    Hacihaliloglu, Ilker
    MEDICAL IMAGE LEARNING WITH LIMITED AND NOISY DATA (MILLAND 2022), 2022, 13559 : 76 - 85
  • [35] SMG: self-supervised masked graph learning for cancer gene identification
    Cui, Yan
    Wang, Zhikang
    Wang, Xiaoyu
    Zhang, Yiwen
    Zhang, Ying
    Pan, Tong
    Zhang, Zhe
    Li, Shanshan
    Guo, Yuming
    Akutsu, Tatsuya
    Song, Jiangning
    BRIEFINGS IN BIOINFORMATICS, 2023, 24 (06)
  • [36] Self-Supervised Monocular Depth Estimation Using Hybrid Transformer Encoder
    Hwang, Seung-Jun
    Park, Sung-Jun
    Baek, Joong-Hwan
    Kim, Byungkyu
    IEEE SENSORS JOURNAL, 2022, 22 (19) : 18762 - 18770
  • [37] TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech
    Liu, Andy T.
    Li, Shang-Wen
    Lee, Hung-yi
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 2351 - 2366
  • [38] A Recommendation Algorithm Based on a Self-supervised Learning Pretrain Transformer
    Xu, Yu-Hao
    Wang, Zhen-Hai
    Wang, Zhi-Ru
    Fan, Rong
    Wang, Xing
    NEURAL PROCESSING LETTERS, 2023, 55 (04) : 4481 - 4497
  • [39] Classification of Polyps in Endoscopic Images Using Self-Supervised Structured Learning
    Huang, Qi-Xian
    Lin, Guo-Shiang
    Sun, Hung-Min
    IEEE ACCESS, 2023, 11 : 50025 - 50037
  • [40] Hierarchically Self-supervised Transformer for Human Skeleton Representation Learning
    Chen, Yuxiao
    Zhao, Long
    Yuan, Jianbo
    Tian, Yu
    Xia, Zhaoyang
    Geng, Shijie
    Han, Ligong
    Metaxas, Dimitris N.
    COMPUTER VISION, ECCV 2022, PT XXVI, 2022, 13686 : 185 - 202