Exploring the Effect of Dataset Diversity in Self-supervised Learning for Surgical Computer Vision

被引:0
|
作者
Jaspers, Tim J. M. [1 ]
de Jonker, Ronald L. P. D. [2 ]
Al Khalil, Yasmina [2 ]
Zeelenberg, Tijn [1 ]
Kusters, Carolus H. J. [1 ]
Li, Yiping [2 ]
van Jaarsveld, Romy C. [3 ]
Bakker, Franciscus H. A. [4 ,5 ]
Ruurda, Jelle P. [3 ]
Brinkman, Willem M. [4 ]
De With, Peter H. N. [1 ]
van der Sommen, Fons [1 ]
机构
[1] Eindhoven Univ Technol, Dept Elect Engn Video Coding & Architectures, Eindhoven, Netherlands
[2] Eindhoven Univ Technol, Dept Biomed Engn, Med Image Anal, Eindhoven, Netherlands
[3] Univ Med Ctr Utrecht, Dept Surg, Utrecht, Netherlands
[4] Univ Med Ctr Utrecht, Dept Oncol Urol, Utrecht, Netherlands
[5] Catharina Hosp, Dept Urol, Eindhoven, Netherlands
来源
DATA ENGINEERING IN MEDICAL IMAGING, DEMI 2024 | 2025年 / 15265卷
关键词
Self-supervised learning; Surgical computer vision; Transfer learning; Data diversity; RECOGNITION;
D O I
10.1007/978-3-031-73748-0_5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Over the past decade, computer vision applications in minimally invasive surgery have rapidly increased. Despite this growth, the impact of surgical computer vision remains limited compared to other medical fields like pathology and radiology, primarily due to the scarcity of representative annotated data. Whereas transfer learning from large annotated datasets such as ImageNet has been conventionally the norm to achieve high-performing models, recent advancements in self-supervised learning (SSL) have demonstrated superior performance. In medical image analysis, in-domain SSL pretraining has already been shown to outperform ImageNet-based initialization. Although unlabeled data in the field of surgical computer vision is abundant, the diversity within this data is limited. This study investigates the role of dataset diversity in SSL for surgical computer vision, comparing procedure-specific datasets against a more heterogeneous general surgical dataset across three different downstream surgical applications. The obtained results show that using solely procedure-specific data can lead to substantial improvements of 13.8%, 9.5%, and 36.8% compared to ImageNet pretraining. However, extending this data with more heterogeneous surgical data further increases performance by an additional 5.0%, 5.2%, and 2.5%, suggesting that increasing diversity within SSL data is beneficial for model performance. The code and pretrained model weights are made publicly available at https://github.com/TimJaspers0801/SurgeNet.
引用
收藏
页码:43 / 53
页数:11
相关论文
共 50 条
  • [21] Quantum self-supervised learning
    Jaderberg, B.
    Anderson, L. W.
    Xie, W.
    Albanie, S.
    Kiffner, M.
    Jaksch, D.
    QUANTUM SCIENCE AND TECHNOLOGY, 2022, 7 (03):
  • [22] SELF-SUPERVISED LEARNING FOR INFANT CRY ANALYSIS
    Gorin, Arsenii
    Subakan, Cem
    Abdoli, Sajjad
    Wang, Junhao
    Latremouille, Samantha
    Onu, Charles
    2023 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW, 2023,
  • [23] EXPLORING THE INTEGRATION OF SPEECH SEPARATION AND RECOGNITION WITH SELF-SUPERVISED LEARNING REPRESENTATION
    Masuyama, Yoshiki
    Chang, Xuankai
    Zhang, Wangyou
    Cornell, Samuele
    Wang, Zhong-Qiu
    Ono, Nobutaka
    Qian, Yanmin
    Watanabe, Shinji
    2023 IEEE WORKSHOP ON APPLICATIONS OF SIGNAL PROCESSING TO AUDIO AND ACOUSTICS, WASPAA, 2023,
  • [24] Self-supervised learning for Environmental Sound Classification
    Tripathi, Achyut Mani
    Mishra, Aakansha
    APPLIED ACOUSTICS, 2021, 182
  • [25] Exploring Self-Supervised Graph Learning in Literature-Based Discovery
    Ding, Juncheng
    Jin, Wei
    2021 IEEE 9TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2021), 2021, : 53 - 62
  • [26] Curriculum learning for self-supervised speaker verification
    Heo, Hee-Soo
    Jung, Jee-weon
    Kang, Jingu
    Kwon, Youngki
    Kim, You Jin
    Lee, Bong-Jin
    Chung, Joon Son
    INTERSPEECH 2023, 2023, : 4693 - 4697
  • [27] EXPLORING FEDERATED SELF-SUPERVISED LEARNING FOR GENERAL PURPOSE AUDIO UNDERSTANDING
    Rehman, Yasar Abbas Ur
    Lau, Kin Wai
    Xie, Yuyang
    Ma, Lan
    Shen, Jiajun
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024, 2024, : 335 - 340
  • [28] Self-supervised Vision Transformers for Writer Retrieval
    Raven, Tim
    Matei, Arthur
    Fink, Gernot A.
    DOCUMENT ANALYSIS AND RECOGNITION-ICDAR 2024, PT II, 2024, 14805 : 380 - 396
  • [29] Self-Supervised Vision Transformers for Malware Detection
    Seneviratne, Sachith
    Shariffdeen, Ridwan
    Rasnayaka, Sanka
    Kasthuriarachchi, Nuran
    IEEE ACCESS, 2022, 10 : 103121 - 103135
  • [30] Gait Recognition with Self-Supervised Learning of Gait Features Based on Vision Transformers
    Pincic, Domagoj
    Susanj, Diego
    Lenac, Kristijan
    SENSORS, 2022, 22 (19)