Clinical Validation of a Deep Learning Algorithm for Detection of Pneumonia on Chest Radiographs in Emergency Department Patients with Acute Febrile Respiratory Illness

被引:34
作者
Kim, Jae Hyun [1 ]
Kim, Jin Young [1 ]
Kim, Gun Ha [1 ]
Kang, Donghoon [2 ]
Kim, In Jung [2 ]
Seo, Jeongkuk [2 ]
Andrews, Jason R. [3 ]
Park, Chang Min [4 ,5 ]
机构
[1] Armed Forces Goyang Hosp, Dept Radiol, 215 Hyeeum Ro, Goyang Si 10271, Gyeonggi Do, South Korea
[2] Armed Forces Goyang Hosp, Dept Internal Med, 215 Hyeeum Ro, Goyang Si 10271, Gyeonggi Do, South Korea
[3] Stanford Univ, Div Infect Dis & Geog Med, Sch Med, 291 Campus Dr, Stanford, CA 94305 USA
[4] Seoul Natl Univ, Dept Radiol, Coll Med, 101 Daehak Ro, Seoul 03080, South Korea
[5] Seoul Natl Univ, Inst Radiat Med, Coll Med, 101 Daehak Ro, Seoul 03080, South Korea
关键词
acute febrile respiratory illness; emergency department; chest radiograph; artificial intelligence; deep learning algorithm; DIABETIC-RETINOPATHY;
D O I
10.3390/jcm9061981
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Early identification of pneumonia is essential in patients with acute febrile respiratory illness (FRI). We evaluated the performance and added value of a commercial deep learning (DL) algorithm in detecting pneumonia on chest radiographs (CRs) of patients visiting the emergency department (ED) with acute FRI. This single-centre, retrospective study included 377 consecutive patients who visited the ED and the resulting 387 CRs in August 2018-January 2019. The performance of a DL algorithm in detection of pneumonia on CRs was evaluated based on area under the receiver operating characteristics (AUROC) curves, sensitivity, specificity, negative predictive values (NPVs), and positive predictive values (PPVs). Three ED physicians independently reviewed CRs with observer performance test to detect pneumonia, which was re-evaluated with the algorithm eight weeks later. AUROC, sensitivity, and specificity measurements were compared between "DL algorithm" vs. "physicians-only" and between "physicians-only" vs. "physicians aided with the algorithm". Among 377 patients, 83 (22.0%) had pneumonia. AUROC, sensitivity, specificity, PPV, and NPV of the algorithm for detection of pneumonia on CRs were 0.861, 58.3%, 94.4%, 74.2%, and 89.1%, respectively. For the detection of 'visible pneumonia on CR' (60 CRs from 59 patients), AUROC, sensitivity, specificity, PPV, and NPV were 0.940, 81.7%, 94.4%, 74.2%, and 96.3%, respectively. In the observer performance test, the algorithm performed better than the physicians for pneumonia (AUROC, 0.861 vs. 0.788,p= 0.017; specificity, 94.4% vs. 88.7%,p< 0.0001) and visible pneumonia (AUROC, 0.940 vs. 0.871,p= 0.007; sensitivity, 81.7% vs. 73.9%,p= 0.034; specificity, 94.4% vs. 88.7%,p< 0.0001). Detection of pneumonia (sensitivity, 82.2% vs. 53.2%,p = 0.008; specificity, 98.1% vs. 88.7%;p< 0.0001) and 'visible pneumonia' (sensitivity, 82.2% vs. 73.9%,p= 0.014; specificity, 98.1% vs. 88.7%,p< 0.0001) significantly improved when the algorithm was used by the physicians. Mean reading time for the physicians decreased from 165 to 101 min with the assistance of the algorithm. Thus, the DL algorithm showed a better diagnosis of pneumonia, particularly visible pneumonia on CR, and improved diagnosis by ED physicians in patients with acute FRI.
引用
收藏
页码:1 / 11
页数:11
相关论文
共 18 条
[1]   Accuracy of chest radiograph interpretation by emergency physicians [J].
Al Aseri Z. .
Emergency Radiology, 2009, 16 (2) :111-114
[2]   Diagnostic Assessment of Deep Learning Algorithms for Detection of Lymph Node Metastases in Women With Breast Cancer [J].
Bejnordi, Babak Ehteshami ;
Veta, Mitko ;
van Diest, Paul Johannes ;
van Ginneken, Bram ;
Karssemeijer, Nico ;
Litjens, Geert ;
van der Laak, Jeroen A. W. M. .
JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 2017, 318 (22) :2199-2210
[3]   Development and Validation of a Deep Learning System for Staging Liver Fibrosis by Using Contrast Agent-enhanced CT Images in the Liver [J].
Choi, Kyu Jin ;
Jang, Jong Keon ;
Lee, Seung Soo ;
Sung, Yu Sub ;
Shim, Woo Hyun ;
Kim, Ho Sung ;
Yun, Jessica ;
Choi, Jin-Young ;
Lee, Yedaun ;
Kang, Bo-Kyeong ;
Kim, Jin Hee ;
Kim, So Yeon ;
Yu, Eun Sil .
RADIOLOGY, 2018, 289 (03) :688-697
[4]   COMPARING THE AREAS UNDER 2 OR MORE CORRELATED RECEIVER OPERATING CHARACTERISTIC CURVES - A NONPARAMETRIC APPROACH [J].
DELONG, ER ;
DELONG, DM ;
CLARKEPEARSON, DI .
BIOMETRICS, 1988, 44 (03) :837-845
[5]   Interpretation of emergency department radiographs: A comparison of emergency medicine physicians with radiologists, residents with faculty, and film with digital display [J].
Eng, J ;
Mysko, WK ;
Weller, GER ;
Renard, R ;
Gitlin, JN ;
Bluemke, DA ;
Magid, D ;
Kelen, GD ;
Scott, WW .
AMERICAN JOURNAL OF ROENTGENOLOGY, 2000, 175 (05) :1233-1238
[6]  
Expert Panel on Thoracic Imaging:, 2018, J Am Coll Radiol, V15, pS240, DOI 10.1016/j.jacr.2018.09.012
[7]  
Ferkol Thomas, 2014, Ann Am Thorac Soc, V11, P404, DOI 10.1513/AnnalsATS.201311-405PS
[8]   Chest radiographs in the emergency department: is the radiologist really necessary? [J].
Gatt, ME ;
Spectre, G ;
Paltiel, O ;
Hiller, N ;
Stalnikowicz, R .
POSTGRADUATE MEDICAL JOURNAL, 2003, 79 (930) :214-217
[9]   Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs [J].
Gulshan, Varun ;
Peng, Lily ;
Coram, Marc ;
Stumpe, Martin C. ;
Wu, Derek ;
Narayanaswamy, Arunachalam ;
Venugopalan, Subhashini ;
Widner, Kasumi ;
Madams, Tom ;
Cuadros, Jorge ;
Kim, Ramasamy ;
Raman, Rajiv ;
Nelson, Philip C. ;
Mega, Jessica L. ;
Webster, R. .
JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 2016, 316 (22) :2402-2410
[10]   CHARACTERISTICS AND TRENDS OF EMERGENCY DEPARTMENT VISITS IN THE UNITED STATES (2010-2014) [J].
Hooker, Edmond A. ;
Mallow, Peter J. ;
Oglesby, Michelle M. .
JOURNAL OF EMERGENCY MEDICINE, 2019, 56 (03) :344-351