A Dataset and Benchmarks for Segmentation and Recognition of Gestures in Robotic Surgery

被引:179
作者
Ahmidi, Narges [1 ]
Tao, Lingling [2 ]
Sefati, Shahin [2 ]
Gao, Yixin [1 ]
Lea, Colin [1 ]
Haro, Benjamin Bejar [2 ]
Zappella, Luca [2 ]
Khudanpur, Sanjeev [3 ]
Vidal, Rene [2 ]
Hager, Gregory D. [1 ]
机构
[1] Johns Hopkins Univ, Dept Comp Sci, Baltimore, MD 21218 USA
[2] Johns Hopkins Univ, Dept Biomed Engn, Ctr Imaging Sci, Baltimore, MD USA
[3] Johns Hopkins Univ, Dept Elect & Comp Engn, Baltimore, MD 21218 USA
基金
欧洲研究理事会; 美国国家科学基金会;
关键词
Activity recognition; benchmark robotic dataset; kinematics and video; surgical motion; OBJECTIVE STRUCTURED ASSESSMENT; BINET-CAUCHY KERNELS; DYNAMICAL-SYSTEMS; SURGICAL SKILL; MOTION; CLASSIFICATION; OSATS; TOOL;
D O I
10.1109/TBME.2016.2647680
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Objective: State-of-the-art techniques for surgical data analysis report promising results for automated skill assessment and action recognition. The contributions of many of these techniques, however, are limited to studyspecific data and validation metrics, making assessment of progress across the field extremely challenging. Methods: In this paper, we address two major problems for surgical data analysis: First, lack of uniform-shared datasets and benchmarks, and second, lack of consistent validation processes. We address the former by presenting the JHU-ISI Gesture and Skill Assessment Working Set (JIGSAWS), a public dataset that we have created to support comparative research benchmarking. JIGSAWS contains synchronized video and kinematic data from multiple performances of robotic surgical tasks by operators of varying skill. We address the latter by presenting a well-documented evaluation methodology and reporting results for six techniques for automated segmentation and classification of time-series data on JIGSAWS. These techniques comprise four temporal approaches for joint segmentation and classification: hidden Markov model, sparse hidden Markov model (HMM), Markov semi-Markov conditional random field, and skip-chain conditional random field; and two feature-based ones that aim to classify fixed segments: bag of spatiotemporal features and linear dynamical systems. Results: Most methods recognize gesture activities with approximately 80% overall accuracy under both leave-one-super-trial-out and leaveone- user-out cross-validation settings. Conclusion: Current methods show promising results on this shared dataset, but room for significant progress remains, particularly for consistent prediction of gesture activities across different surgeons. Significance: The results reported in this paper provide the first systematic and uniform evaluation of surgical activity recognition techniques on the benchmark database.
引用
收藏
页码:2025 / 2041
页数:17
相关论文
共 76 条
  • [1] Afsari B, 2012, PROC CVPR IEEE, P2208, DOI 10.1109/CVPR.2012.6247929
  • [2] K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation
    Aharon, Michal
    Elad, Michael
    Bruckstein, Alfred
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2006, 54 (11) : 4311 - 4322
  • [3] Automated objective surgical skill assessment in the operating room from unstructured tool motion in septoplasty
    Ahmidi, Narges
    Poddar, Piyush
    Jones, Jonathan D.
    Vedula, S. Swaroop
    Ishii, Lisa
    Hager, Gregory D.
    Ishii, Masaru
    [J]. INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2015, 10 (06) : 981 - 991
  • [4] Ahmidi N, 2013, LECT NOTES COMPUT SC, V8149, P26, DOI 10.1007/978-3-642-40811-3_4
  • [5] [Anonymous], 2010, THESIS
  • [6] [Anonymous], 2009, P BRIT MACH VIS C
  • [7] Bass BL, 2007, AM SURGEON, V73, P109
  • [8] Eigenfaces vs. Fisherfaces: Recognition using class specific linear projection
    Belhumeur, PN
    Hespanha, JP
    Kriegman, DJ
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1997, 19 (07) : 711 - 720
  • [9] Workflow mining for visualization and analysis of surgeries
    Blum, Tobias
    Padoy, Nicolas
    Feussner, Hubertus
    Navab, Nassir
    [J]. INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2008, 3 (05) : 379 - 386
  • [10] Chaudhry R, 2009, PROC CVPR IEEE, P1932, DOI 10.1109/CVPRW.2009.5206821