Crowdsourcing: a valid alternative to expert evaluation of robotic surgery skills

被引:49
作者
Polin, Michael R. [1 ]
Siddiqui, Nazema Y. [1 ]
Comstock, Bryan A. [2 ]
Hesham, Helai [5 ]
Brown, Casey [5 ]
Lendvay, Thomas S. [3 ]
Martino, Martin A. [4 ,5 ]
机构
[1] Duke Univ, Dept Obstet & Gynecol, Div Urogynecol, Durham, NC USA
[2] Univ Washington, Dept Biostat, Seattle, WA 98195 USA
[3] Univ Washington, Dept Urol, Div Pediat Urol, Seattle, WA 98195 USA
[4] Lehigh Valley Hlth Network, Div Gynecol Oncol, Allentown, PA USA
[5] Lehigh Valley Hlth Network, Dept Obstet & Gynecol, Allentown, PA USA
关键词
crowdsourcing; robotic surgery; simulation; surgical training; CROWD-SOURCED ASSESSMENT; OBJECTIVE STRUCTURED ASSESSMENT; TECHNICAL SKILLS; SURGICAL SKILL; LEARNING-CURVE; PERFORMANCE;
D O I
10.1016/j.ajog.2016.06.033
中图分类号
R71 [妇产科学];
学科分类号
100211 ;
摘要
BACKGROUND: Robotic-assisted gynecologic surgery is common, but requires unique training. A validated assessment tool for evaluating trainees' robotic surgery skills is Robotic-Objective Structured Assessments of Technical Skills. OBJECTIVE: We sought to assess whether crowdsourcing can be used as an alternative to expert surgical evaluators in scoring Robotic-Objective Structured Assessments of Technical Skills. STUDY DESIGN: The Robotic Training Network produced the Robotic-Objective Structured Assessments of Technical Skills, which evaluate trainees across 5 dry lab robotic surgical drills. Robotic-Objective Structured Assessments of Technical Skills were previously validated in a study of 105 participants, where dry lab surgical drills were recorded, de-identified, and scored by 3 expert surgeons using the Robotic-Objective Structured Assessments of Technical Skills checklist. Our methods-comparison study uses these previously obtained recordings and expert surgeon scores. Mean scores per participant from each drill were separated into quartiles. Crowdworkers were trained and calibrated on Robotic-Objective Structured Assessments of Technical Skills scoring using a representative recording of a skilled and novice surgeon. Following this, 3 recordings from each scoring quartile for each drill were randomly selected. Crowdworkers evaluated the randomly selected recordings using Robotic-Objective Structured Assessments of Technical Skills. Linear mixed effects models were used to derive mean crowdsourced ratings for each drill. Pearson correlation coefficients were calculated to assess the correlation between crowdsourced and expert surgeons' ratings. RESULTS: In all, 448 crowdworkers reviewed videos from 60 dry lab drills, and completed a total of 2517 Robotic-Objective Structured Assessments of Technical Skills assessments within 16 hours. Crowd-sourced Robotic-Objective Structured Assessments of Technical Skills ratings were highly correlated with expert surgeon ratings across each of the 5 dry lab drills (r ranging from 0.75-0.91). CONCLUSION: Crowdsourced assessments of recorded dry lab surgical drills using a validated assessment tool are a rapid and suitable alternative to expert surgeon evaluation.
引用
收藏
页数:7
相关论文
共 20 条
[1]   Crowdsourcing Malaria Parasite Quantification: An Online Game for Analyzing Images of Infected Thick Blood Smears [J].
Angel Luengo-Oroz, Miguel ;
Arranz, Asier ;
Frean, John .
JOURNAL OF MEDICAL INTERNET RESEARCH, 2012, 14 (06) :207-219
[2]   Surgical Skill and Complication Rates after Bariatric Surgery [J].
Birkmeyer, John D. ;
Finks, Jonathan F. ;
O'Reilly, Amanda ;
Oerline, Mary ;
Carlin, Arthur M. ;
Nunn, Andre R. ;
Dimick, Justin ;
Banerjee, Mousumi ;
Birkmeyer, Nancy J. O. .
NEW ENGLAND JOURNAL OF MEDICINE, 2013, 369 (15) :1434-1442
[3]   Crowd-Sourced Assessment of Technical Skills: a novel method to evaluate surgical performance [J].
Chen, Carolyn ;
White, Lee ;
Kowalewski, Timothy ;
Aggarwal, Rajesh ;
Lintott, Chris ;
Comstock, Bryan ;
Kuksenok, Katie ;
Aragon, Cecilia ;
Holst, Daniel ;
Lendvay, Thomas .
JOURNAL OF SURGICAL RESEARCH, 2014, 187 (01) :65-71
[4]  
Cicchetti DV., 1994, PSYCHOL ASSESSMENTS, V6, P284, DOI [10.1037/1040-3590.6.4.284, DOI 10.1037/1040-3590.6.4.284]
[5]   Preliminary results from a crowdsourcing experiment in immunohistochemistry [J].
Della Mea, Vincenzo ;
Maddalena, Eddy ;
Mizzaro, Stefano ;
Machin, Piernicola ;
Beltrami, Carlo A. .
DIAGNOSTIC PATHOLOGY, 2014, 9
[6]   Measuring to Improve: Peer and Crowd-sourced Assessments of Technical Skill with Robot-assisted Radical Prostatectomy [J].
Ghani, Khurshid R. ;
Miller, David C. ;
Linsell, Susan ;
Brachulis, Andrew ;
Lane, Brian ;
Sarle, Richard ;
Dalela, Deepansh ;
Menon, Mani ;
Comstock, Bryan ;
Lendvay, Thomas S. ;
Montie, James ;
Peabody, James O. .
EUROPEAN UROLOGY, 2016, 69 (04) :547-550
[7]   Design, Analysis, and Interpretation of Method-Comparison Studies [J].
Hanneman, Sandra K. .
AACN ADVANCED CRITICAL CARE, 2008, 19 (02) :223-234
[8]   Crowd-Sourced Assessment of Technical Skills: Differentiating Animate Surgical Skill Through the Wisdom of Crowds [J].
Holst, Daniel ;
Kowalewski, Timothy M. ;
White, Lee W. ;
Brand, Timothy C. ;
Harper, Jonathan D. ;
Sorensen, Mathew D. ;
Truong, Mireille ;
Simpson, Khara ;
Tanaka, Alyssa ;
Smith, Roger ;
Lendvay, Thomas S. .
JOURNAL OF ENDOUROLOGY, 2015, 29 (10) :1183-1188
[9]   Crowd-Sourced Assessment of Technical Skills: An Adjunct to Urology Resident Surgical Simulation Training [J].
Holst, Daniel ;
Kowalewski, Timothy M. ;
White, Lee W. ;
Brand, Timothy C. ;
Harper, Jonathan D. ;
Sorenson, Mathew D. ;
Kirsch, Sarah ;
Lendvay, Thomas S. .
JOURNAL OF ENDOUROLOGY, 2015, 29 (05) :604-609
[10]   RANDOM-EFFECTS MODELS FOR LONGITUDINAL DATA [J].
LAIRD, NM ;
WARE, JH .
BIOMETRICS, 1982, 38 (04) :963-974