Can artificial intelligence pass the Fellowship of the Royal College of Radiologists examination? Multi-reader diagnostic accuracy study

被引:22
作者
Shelmerdine, Susan Cheng [1 ,2 ,3 ,4 ]
Martin, Helena [4 ]
Shirodhar, Kapil [5 ]
Shamshuddin, Sameer [5 ]
Weir-McCall, Jonathan Richard [6 ,7 ]
机构
[1] Great Ormond St Hosp Sick Children, Dept Clin Radiol, London, England
[2] Great Ormond St Hosp Sick Children, UCL Great Ormond St Inst Child Hlth, London, England
[3] NIHR Great Ormond St Hosp Biomed Res Ctr, London, England
[4] St George Hosp, Dept Clin Radiol, London, England
[5] Univ Hosp Morecambe Bay NHS Trust, Royal Lancaster Infirm, Dept Radiol, Lancaster, England
[6] Univ Cambridge, Sch Clin Med, Cambridge, England
[7] Royal Papworth Hosp, Dept Radiol, Cambridge, England
来源
BMJ-BRITISH MEDICAL JOURNAL | 2022年 / 379卷
关键词
D O I
10.1136/bmj-2022-072826
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
OBJECTIVE To determine whether an artificial intelligence candidate could pass the rapid (radiographic) reporting component of the Fellowship of the Royal College of Radiologists (FRCR) examination. DESIGN Prospective multi-reader diagnostic accuracy study. SETTING United Kingdom. PARTICIPANTS One artificial intelligence candidate (Smarturgences, Milvue) and 26 radiologists who had passed the FRCR examination in the preceding 12 months. MAIN OUTCOME MEASURES Accuracy and pass rate of the artificial intelligence compared with radiologists across 10 mock FRCR rapid reporting examinations (each examination containing 30 radiographs, requiring 90% accuracy rate to pass). RESULTS When non-interpretable images were excluded from the analysis, the artificial intelligence candidate achieved an average overall accuracy of 79.5% (95% confidence interval 74.1% to 84.3%) and passed two of 10 mock FRCR examinations. The average radiologist achieved an average accuracy of 84.8% (76.1-91.9%) and passed four of 10 mock examinations. The sensitivity for the artificial intelligence was 83.6% (95% confidence interval 76.2% to 89.4%) and the specificity was 75.2% (66.7% to 82.5%), compared with summary estimates across all radiologists of 84.1% (81.0% to 87.0%) and 87.3% (85.0% to 89.3%). Across 148/300 radiographs that were correctly interpreted by >90% of radiologists, the artificial intelligence candidate was incorrect in 14/148 (9%). In 20/300 radiographs that most (>50%) radiologists interpreted incorrectly, the artificial intelligence candidate was correct in 10/20 (50%). Most imaging pitfalls related to interpretation of musculoskeletal rather than chest radiographs. CONCLUSIONS When special dispensation for the artificial intelligence candidate was provided (that is, exclusion of non-interpretable images), the artificial intelligence candidate was able to pass two of 10 mock examinations. Potential exists for the artificial intelligence candidate to improve its radiographic interpretation skills by focusing on musculoskeletal cases and learning to interpret radiographs of the axial skeleton and abdomen that are currently considered "non-interpretable."
引用
收藏
页数:12
相关论文
共 19 条
[1]   Diagnostic accuracy of deep learning in medical imaging: a systematic review and meta-analysis [J].
Aggarwal, Ravi ;
Sounderajah, Viknesh ;
Martin, Guy ;
Ting, Daniel S. W. ;
Karthikesalingam, Alan ;
King, Dominic ;
Ashrafian, Hutan ;
Darzi, Ara .
NPJ DIGITAL MEDICINE, 2021, 4 (01)
[2]   Fitting Linear Mixed-Effects Models Using lme4 [J].
Bates, Douglas ;
Maechler, Martin ;
Bolker, Benjamin M. ;
Walker, Steven C. .
JOURNAL OF STATISTICAL SOFTWARE, 2015, 67 (01) :1-48
[3]   The Fellowship of the Royal College of Radiologists (FRCR) examination: a review of the evidence [J].
Booth, T. C. ;
Martins, R. de P. M. ;
McKnight, L. ;
Courtney, K. ;
Malliwal, R. .
CLINICAL RADIOLOGY, 2018, 73 (12) :992-998
[4]   Machine Learning: Discovering the Future of Medical Imaging [J].
Erickson, Bradley J. .
JOURNAL OF DIGITAL IMAGING, 2017, 30 (04) :391-391
[5]  
Grande Challenge, AI RAD IMPL GUID MIL
[6]   Improving Radiographic Fracture Recognition Performance and Efficiency Using Artificial Intelligence [J].
Guermazi, Ali ;
Tannoury, Chadi ;
Kompel, Andrew J. ;
Murakami, Akira M. ;
Ducarouge, Alexis ;
Gillibert, Andre ;
Li, Xinning ;
Tournier, Antoine ;
Lahoud, Youmna ;
Jarraya, Mohamed ;
Lacave, Elise ;
Rahimi, Hamza ;
Pourchot, Alois ;
Parisien, Robert L. ;
Merritt, Alexander C. ;
Comeau, Douglas ;
Regnard, Nor-Eddine ;
Hayashi, Daichi .
RADIOLOGY, 2022, 302 (03) :627-636
[7]   Radiology training in United Kingdom: current status [J].
Kassamali, Rahil H. ;
Hoey, Edward T. D. .
QUANTITATIVE IMAGING IN MEDICINE AND SURGERY, 2014, 4 (06) :447-448
[8]   Artificial Intelligence in Fracture Detection: A Systematic Review and Meta-Analysis [J].
Kuo, Rachel Y. L. ;
Harrison, Conrad ;
Curran, Terry-Ann ;
Jones, Benjamin ;
Freethy, Alexander ;
Cussons, David ;
Stewart, Max ;
Collins, Gary S. ;
Furniss, Dominic .
RADIOLOGY, 2022, 304 (01) :50-62
[9]  
Mukherjee S. A.I., 2017, NEW YORKER
[10]   Artificial intelligence: Who is responsible for the diagnosis? [J].
Neri, Emanuele ;
Coppola, Francesca ;
Miele, Vittorio ;
Bibbolino, Corrado ;
Grassi, Roberto .
RADIOLOGIA MEDICA, 2020, 125 (06) :517-521