Comparison of emergency medicine specialist, cardiologist, and chat-GPT in electrocardiography assessment

被引:38
作者
Gunay, Serkan [1 ]
Ozturk, Ahmet [1 ]
Ozerol, Hakan [2 ]
Yigit, Yavuz [3 ]
Erenler, Ali Kemal [1 ]
机构
[1] Hitit Univ, Erol Olcok Educ & Res Hosp, Dept Emergency Med, Inonu Cd 176, TR-19040 Corum, Turkiye
[2] Gaziantep City Hosp, Dept Emergency Med, Gaziantep, Turkiye
[3] Hamad Gen Hosp, Dept Emergency Med, Hamad Med Corp, Doha, Qatar
关键词
ChatGPT; Electrocardiography; Emergency medicine; Cardiology;
D O I
10.1016/j.ajem.2024.03.017
中图分类号
R4 [临床医学];
学科分类号
1002 ; 100602 ;
摘要
Introduction: ChatGPT, developed by OpenAI, represents the cutting-edge in its field with its latest model, GPT-4. Extensive research is currently being conducted in various domains, including cardiovascular diseases, using ChatGPT. Nevertheless, there is a lack of studies addressing the proficiency of GPT-4 in diagnosing conditions based on Electrocardiography (ECG) data. The goal of this study is to evaluate the diagnostic accuracy of GPT-4 when provided with ECG data, and to compare its performance with that of emergency medicine specialists and cardiologists. Methods: This study has received approval from the Clinical Research Ethics Committee of Hitit University Medical Faculty on August 21, 2023 (decision no: 2023 -91). Drawing on cases from the "150 ECG Cases " book, a total of 40 ECG cases were crafted into multiple-choice questions (comprising 20 everyday and 20 more challenging ECG questions). The participant pool included 12 emergency medicine specialists and 12 cardiology specialists. GPT-4 was administered the questions in a total of 12 separate sessions. The responses from the cardiology physicians, emergency medicine physicians, and GPT-4 were evaluated separately for each of the three groups. Results: In the everyday ECG questions, GPT-4 demonstrated superior performance compared to both the emergency medicine specialists and the cardiology specialists ( p < 0.001, p = 0.001). In the more challenging ECG questions, while Chat-GPT outperformed the emergency medicine specialists (p < 0.001), no significant statistical difference was found between Chat-GPT and the cardiology specialists ( p = 0.190). Upon examining the accuracy of the total ECG questions, Chat-GPT was found to be more successful compared to both the Emergency Medicine Specialists and the cardiologists ( p < 0.001, p = 0.001). Conclusion: Our study has shown that GPT-4 is more successful than emergency medicine specialists in evaluating both everyday and more challenging ECG questions. It performed better compared to cardiologists on everyday questions, but its performance aligned closely with that of the cardiologists as the difficulty of the questions increased. (c) 2024 Elsevier Inc. All rights reserved.
引用
收藏
页码:51 / 60
页数:10
相关论文
共 17 条
[1]   Professionalism in emergency medicine [J].
Adams, J ;
Schmidt, T ;
Sanders, A ;
Larkin, GL ;
Knopp, R .
ACADEMIC EMERGENCY MEDICINE, 1998, 5 (12) :1193-1199
[2]   Usefulness of Machine Learning-Based Detection and Classification of Cardiac Arrhythmias With 12-Lead Electrocardiograms [J].
Chang, Kuan-Cheng ;
Hsieh, Po-Hsin ;
Wu, Mei-Yao ;
Wang, Yu-Chen ;
Chen, Jan-Yow ;
Tsai, Fuu-Jen ;
Shih, Edward S. C. ;
Hwang, Ming-Jing ;
Huang, Tzung-Chi .
CANADIAN JOURNAL OF CARDIOLOGY, 2021, 37 (01) :94-104
[3]  
ChatGPT, 2024, Homepage on the Internet. c2023
[4]   Performance of ChatGPT-4 in answering questions from the Brazilian National Examination for Medical Degree Revalidation [J].
Gobira, Mauro ;
Nakayama, Luis Filipe ;
Moreira, Rodrigo ;
Andrade, Eric ;
Regatieri, Caio Vinicius Saito ;
Belfort Jr, Rubens .
REVISTA DA ASSOCIACAO MEDICA BRASILEIRA, 2023, 69 (10)
[5]  
Hamet P, 2017, METABOLISM, V69, pS36, DOI [10.26738/mjem.2017/mjem26.2018/plh.cre.041017, 10.1016/j.metabol.2017.01.011]
[6]  
Hampton J, 2019, 150 ECG cases
[7]   Assessment of Resident and AI Chatbot Performance on the University of Toronto Family Medicine Residency Progress Test: Comparative Study [J].
Huang, Ryan S. T. ;
Lu, Kevin Jia Qi ;
Meaney, Christopher ;
Kemppainen, Joel ;
Punnett, Angela ;
Leung, Fok-Han .
JMIR MEDICAL EDUCATION, 2023, 9
[8]   Performance of a Convolutional Neural Network and Explainability Technique for 12-Lead Electrocardiogram Interpretation [J].
Hughes, J. Weston ;
Olgin, Jeffrey E. ;
Avram, Robert ;
Abreau, Sean A. ;
Sittler, Taylor ;
Radia, Kaahan ;
Hsia, Henry ;
Walters, Tomos ;
Lee, Byron ;
Gonzalez, Joseph E. ;
Tison, Geoffrey H. .
JAMA CARDIOLOGY, 2021, 6 (11) :1285-1295
[9]   ChatGPT Versus Human Performance on Emergency Medicine Board Preparation Questions [J].
Jarou, Zachary J. ;
Dakka, Ali ;
McGuire, Duncan ;
Bunting, Leonard .
ANNALS OF EMERGENCY MEDICINE, 2024, 83 (01) :87-88
[10]   Performance of ChatGPT on USMLE: Potential for AI-assisted medical education using large language models [J].
Kung, Tiffany H. ;
Cheatham, Morgan ;
Medenilla, Arielle ;
Sillos, Czarina ;
De Leon, Lorie ;
Elepano, Camille ;
Madriaga, Maria ;
Aggabao, Rimel ;
Diaz-Candido, Giezel ;
Maningo, James ;
Tseng, Victor .
PLOS DIGITAL HEALTH, 2023, 2 (02)