MS-DINO: Masked Self-Supervised Distributed Learning Using Vision Transformer

被引:1
|
作者
Park, Sangjoon [1 ,2 ,3 ]
Lee, Ik Jae [4 ]
Kim, Jun Won [4 ]
Ye, Jong Chul [5 ]
机构
[1] Korea Adv Inst Sci & Technol, Dept Bio & Brain Engn, Daejeon 34141, South Korea
[2] Yonsei Univ, Coll Med, Dept Radiat Oncol, Seoul 03722, South Korea
[3] Yonsei Univ, Inst Innovat Digital Healthcare, Seoul 03722, South Korea
[4] Gangnam Severance Hosp, Dept Radiat Oncol, Seoul 06273, South Korea
[5] Korea Adv Inst Sci & Technol, Kim Jaechul Grad Sch AI, Daejeon 34141, South Korea
基金
新加坡国家研究基金会;
关键词
Feature extraction; Task analysis; Biomedical imaging; Privacy; Transformers; Servers; Distance learning; Distributed learning; self-supervised learning; random permutation; vision transformer; privacy protection;
D O I
10.1109/JBHI.2024.3423797
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Despite promising advancements in deep learning in medical domains, challenges still remain owing to data scarcity, compounded by privacy concerns and data ownership disputes. Recent explorations of distributed-learning paradigms, particularly federated learning, have aimed to mitigate these challenges. However, these approaches are often encumbered by substantial communication and computational overhead, and potential vulnerabilities in privacy safeguards. Therefore, we propose a self-supervised masked sampling distillation technique called MS-DINO, tailored to the vision transformer architecture. This approach removes the need for incessant communication and strengthens privacy using a modified encryption mechanism inherent to the vision transformer while minimizing the computational burden on client-side devices. Rigorous evaluations across various tasks confirmed that our method outperforms existing self-supervised distributed learning strategies and fine-tuned baselines.
引用
收藏
页码:6180 / 6192
页数:13
相关论文
共 50 条
  • [1] A Hierarchical Vision Transformer Using Overlapping Patch and Self-Supervised Learning
    Ma, Yaxin
    Li, Ming
    Chang, Jun
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [2] Self-Supervised Pretraining Vision Transformer With Masked Autoencoders for Building Subsurface Model
    Li, Yuanyuan
    Alkhalifah, Tariq
    Huang, Jianping
    Li, Zhenchun
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [3] Self-Supervised Pretraining Vision Transformer With Masked Autoencoders for Building Subsurface Model
    Li, Yuanyuan
    Alkhalifah, Tariq
    Huang, Jianping
    Li, Zhenchun
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [4] MST: Masked Self-Supervised Transformer for Visual Representation
    Li, Zhaowen
    Chen, Zhiyang
    Yang, Fan
    Li, Wei
    Zhu, Yousong
    Zhao, Chaoyang
    Deng, Rui
    Wu, Liwei
    Zhao, Rui
    Tang, Ming
    Wang, Jinqiao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] Positional Label for Self-Supervised Vision Transformer
    Zhang, Zhemin
    Gong, Xun
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 3, 2023, : 3516 - 3524
  • [6] Self-Supervised Masked Convolutional Transformer Block for Anomaly Detection
    Madan, Neelu
    Ristea, Nicolae-Catalin
    Ionescu, Radu Tudor
    Nasrollahi, Kamal
    Khan, Fahad Shahbaz
    Moeslund, Thomas B.
    Shah, Mubarak
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (01) : 525 - 542
  • [7] Pattern Integration and Enhancement Vision Transformer for Self-Supervised Learning in Remote Sensing
    Lu, Kaixuan
    Zhang, Ruiqian
    Huang, Xiao
    Xie, Yuxing
    Ning, Xiaogang
    Zhang, Hanchao
    Yuan, Mengke
    Zhang, Pan
    Wang, Tao
    Liao, Tongkui
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2025, 63
  • [8] Scene Interpretation Method using Transformer and Self-supervised Learning
    Kobayashi, Yuya
    Suzuki, Masahiro
    Matsuo, Yutaka
    Transactions of the Japanese Society for Artificial Intelligence, 2022, 37 (02)
  • [9] Masked Discrimination for Self-supervised Learning on Point Clouds
    Liu, Haotian
    Cai, Mu
    Lee, Yong Jae
    COMPUTER VISION - ECCV 2022, PT II, 2022, 13662 : 657 - 675
  • [10] A Survey on Masked Autoencoder for Visual Self-supervised Learning
    Zhang, Chaoning
    Zhang, Chenshuang
    Song, Junha
    Yi, John Seon Keun
    Kweon, In So
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 6805 - 6813