PhacoTrainer: Deep Learning for Cataract Surgical Videos to Track Surgical Tools

被引:7
|
作者
Yeh, Hsu-Hang [1 ]
Jain, Anjal M. [2 ]
Fox, Olivia [3 ]
Sebov, Kostya [1 ]
Wang, Sophia Y. [1 ,2 ,4 ]
机构
[1] Stanford Univ, Dept Biomed Data Sci, Palo Alto, CA USA
[2] Stanford Univ, Byers Eye Inst, Dept Ophthalmol, Palo Alto, CA USA
[3] Johns Hopkins Univ, Krieger Sch Arts & Sci, Baltimore, MD USA
[4] 2370 Watson Court, Palo Alto, CA 94303 USA
来源
关键词
deep learning; cataract surgery; surgical training; surgical performance; artificial intelligence; SKILL; TIME;
D O I
10.1167/tvst.12.3.23
中图分类号
R77 [眼科学];
学科分类号
100212 ;
摘要
Purpose: The purpose of this study was to build a deep-learning model that automat-ically analyzes cataract surgical videos for the locations of surgical landmarks, and to derive skill-related motion metrics. Methods: The locations of the pupil, limbus, and 8 classes of surgical instruments were identified by a 2-step algorithm: (1) mask segmentation and (2) landmark identifica-tion from the masks. To perform mask segmentation, we trained the YOLACT model on 1156 frames sampled from 268 videos and the public Cataract Dataset for Image Segmentation (CaDIS) dataset. Landmark identification was performed by fitting ellipses or lines to the contours of the masks and deriving locations of interest, including surgi-cal tooltips and the pupil center. Landmark identification was evaluated by the distance between the predicted and true positions in 5853 frames of 10 phacoemulsification video clips. We derived the total path length, maximal speed, and covered area using the tip positions and examined the correlation with human-rated surgical performance.Results: The mean average precision score and intersection-over-union for mask detec-tion were 0.78 and 0.82. The average distance between the predicted and true positions of the pupil center, phaco tip, and second instrument tip was 5.8, 9.1, and 17.1 pixels. The total path length and covered areas of these landmarks were negatively correlated with surgical performance.Conclusions: We developed a deep-learning method to localize key anatomical portions of the eye and cataract surgical tools, which can be used to automatically derive metrics correlated with surgical skill.Translational Relevance: Our system could form the basis of an automated feedback system that helps cataract surgeons evaluate their performance.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] PhacoTrainer: Deep Learning for Cataract Surgical Videos to Track Surgical Tools
    Yeh, Hsu-Hang
    Jain, Anjal M.
    Jallow, Mariama
    Sebov, Kostya
    Wang, Sophia Y.
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2022, 63 (07)
  • [2] PhacoTrainer: Deep Learning for Activity Recognition in Cataract Surgical Videos
    Yeh, Hsu-Hang
    Jain, Anjal
    Fox, Olivia
    Wang, Sophia
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2021, 62 (08)
  • [3] PhacoTrainer: A Multicenter Study of Deep Learning for Activity Recognition in Cataract Surgical Videos
    Yeh, Hsu-Hang
    Jain, Anjal M.
    Fox, Olivia
    Wang, Sophia Y.
    TRANSLATIONAL VISION SCIENCE & TECHNOLOGY, 2021, 10 (13):
  • [4] A contextual detector of surgical tools in laparoscopic videos using deep learning
    Namazi, Babak
    Sankaranarayanan, Ganesh
    Devarajan, Venkat
    SURGICAL ENDOSCOPY AND OTHER INTERVENTIONAL TECHNIQUES, 2022, 36 (01): : 679 - 688
  • [5] A contextual detector of surgical tools in laparoscopic videos using deep learning
    Babak Namazi
    Ganesh Sankaranarayanan
    Venkat Devarajan
    Surgical Endoscopy, 2022, 36 : 679 - 688
  • [6] A deep learning spatial-temporal framework for detecting surgical tools in laparoscopic videos
    Abdulbaki Alshirbaji, Tamer
    Jalal, Nour Aldeen
    Docherty, Paul D.
    Neumuth, Thomas
    Möller, Knut
    Biomedical Signal Processing and Control, 2021, 68
  • [7] A deep learning spatial-temporal framework for detecting surgical tools in laparoscopic videos
    Alshirbaji, Tamer Abdulbaki
    Jalal, Nour Aldeen
    Docherty, Paul D.
    Neumuth, Thomas
    Moeller, Knut
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2021, 68
  • [8] Beyond PhacoTrainer: Deep Learning for Enhanced Trabecular Meshwork Detection in MIGS Videos
    Kara, Su
    Yang, Michael
    Yeh, Hsu-Hang
    Sen, Simmi
    Hwang, Hannah H.
    Wang, Sophia Y.
    TRANSLATIONAL VISION SCIENCE & TECHNOLOGY, 2024, 13 (09):
  • [9] Deep Neural Architecture for Localization and Tracking of Surgical Tools in Cataract Surgery
    Banerjee, Neha
    Sathish, Rachana
    Sheet, Debdoot
    COMPUTER AIDED INTERVENTION AND DIAGNOSTICS IN CLINICAL AND MEDICAL IMAGES, 2019, 31 : 31 - 38
  • [10] Deep learning for surgical phase recognition using endoscopic videos
    Annetje C. P. Guédon
    Senna E. P. Meij
    Karim N. M. M. H. Osman
    Helena A. Kloosterman
    Karlijn J. van Stralen
    Matthijs C. M. Grimbergen
    Quirijn A. J. Eijsbouts
    John J. van den Dobbelsteen
    Andru P. Twinanda
    Surgical Endoscopy, 2021, 35 : 6150 - 6157