Exploring the Effect of Dataset Diversity in Self-supervised Learning for Surgical Computer Vision

被引:0
|
作者
Jaspers, Tim J. M. [1 ]
de Jonker, Ronald L. P. D. [2 ]
Al Khalil, Yasmina [2 ]
Zeelenberg, Tijn [1 ]
Kusters, Carolus H. J. [1 ]
Li, Yiping [2 ]
van Jaarsveld, Romy C. [3 ]
Bakker, Franciscus H. A. [4 ,5 ]
Ruurda, Jelle P. [3 ]
Brinkman, Willem M. [4 ]
De With, Peter H. N. [1 ]
van der Sommen, Fons [1 ]
机构
[1] Eindhoven Univ Technol, Dept Elect Engn Video Coding & Architectures, Eindhoven, Netherlands
[2] Eindhoven Univ Technol, Dept Biomed Engn, Med Image Anal, Eindhoven, Netherlands
[3] Univ Med Ctr Utrecht, Dept Surg, Utrecht, Netherlands
[4] Univ Med Ctr Utrecht, Dept Oncol Urol, Utrecht, Netherlands
[5] Catharina Hosp, Dept Urol, Eindhoven, Netherlands
来源
DATA ENGINEERING IN MEDICAL IMAGING, DEMI 2024 | 2025年 / 15265卷
关键词
Self-supervised learning; Surgical computer vision; Transfer learning; Data diversity; RECOGNITION;
D O I
10.1007/978-3-031-73748-0_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Over the past decade, computer vision applications in minimally invasive surgery have rapidly increased. Despite this growth, the impact of surgical computer vision remains limited compared to other medical fields like pathology and radiology, primarily due to the scarcity of representative annotated data. Whereas transfer learning from large annotated datasets such as ImageNet has been conventionally the norm to achieve high-performing models, recent advancements in self-supervised learning (SSL) have demonstrated superior performance. In medical image analysis, in-domain SSL pretraining has already been shown to outperform ImageNet-based initialization. Although unlabeled data in the field of surgical computer vision is abundant, the diversity within this data is limited. This study investigates the role of dataset diversity in SSL for surgical computer vision, comparing procedure-specific datasets against a more heterogeneous general surgical dataset across three different downstream surgical applications. The obtained results show that using solely procedure-specific data can lead to substantial improvements of 13.8%, 9.5%, and 36.8% compared to ImageNet pretraining. However, extending this data with more heterogeneous surgical data further increases performance by an additional 5.0%, 5.2%, and 2.5%, suggesting that increasing diversity within SSL data is beneficial for model performance. The code and pretrained model weights are made publicly available at https://github.com/TimJaspers0801/SurgeNet.
引用
收藏
页码:43 / 53
页数:11
相关论文
共 50 条
  • [11] Exploring Self-Supervised Learning for 3D Point Cloud Registration
    Yuan, Mingzhi
    Huang, Qiao
    Shen, Ao
    Huang, Xiaoshui
    Wang, Manning
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (01): : 25 - 31
  • [12] Synergistic Self-supervised and Quantization Learning
    Cao, Yun-Hao
    Sun, Peiqin
    Huang, Yechang
    Wu, Jianxin
    Zhou, Shuchang
    COMPUTER VISION - ECCV 2022, PT XXX, 2022, 13690 : 587 - 604
  • [13] A Survey on Contrastive Self-Supervised Learning
    Jaiswal, Ashish
    Babu, Ashwin Ramesh
    Zadeh, Mohammad Zaki
    Banerjee, Debapriya
    Makedon, Fillia
    TECHNOLOGIES, 2021, 9 (01)
  • [14] Y Modular Self-Supervised Learning for Hand Surgical Diagnosis
    Dechaumet, Leo
    Bennani, Younes
    Karkazan, Joseph
    Barbara, Abir
    Dacheux, Charles
    Gregory, Thomas
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [15] Gated Self-supervised Learning for Improving Supervised Learning
    Fuadi, Erland Hillman
    Ruslim, Aristo Renaldo
    Wardhana, Putu Wahyu Kusuma
    Yudistira, Novanto
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 611 - 615
  • [16] A Hierarchical Vision Transformer Using Overlapping Patch and Self-Supervised Learning
    Ma, Yaxin
    Li, Ming
    Chang, Jun
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [17] Self-Supervised Learning for Recommendation
    Huang, Chao
    Xia, Lianghao
    Wang, Xiang
    He, Xiangnan
    Yin, Dawei
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 5136 - 5139
  • [18] Longitudinal self-supervised learning
    Zhao, Qingyu
    Liu, Zixuan
    Adeli, Ehsan
    Pohl, Kilian M.
    MEDICAL IMAGE ANALYSIS, 2021, 71
  • [19] Reachability Embeddings: Scalable Self-Supervised Representation Learning from Mobility Trajectories for Multimodal Geospatial Computer Vision
    Ganguli, Swetava
    Iyer, C. V. Krishnakumar
    Pandey, Vipul
    2022 23RD IEEE INTERNATIONAL CONFERENCE ON MOBILE DATA MANAGEMENT (MDM 2022), 2022, : 44 - 53
  • [20] Self-Supervised Speech Representation Learning: A Review
    Mohamed, Abdelrahman
    Lee, Hung-yi
    Borgholt, Lasse
    Havtorn, Jakob D.
    Edin, Joakim
    Igel, Christian
    Kirchhoff, Katrin
    Li, Shang-Wen
    Livescu, Karen
    Maaloe, Lars
    Sainath, Tara N.
    Watanabe, Shinji
    IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (06) : 1179 - 1210