Evaluation of responses to cardiac imaging questions by the artificial intelligence large language model ChatGPT

被引:7
作者
Monroe, Cynthia L. [1 ]
Abdelhafez, Yasser G. [2 ]
Atsina, Kwame [3 ]
Aman, Edris [3 ]
Nardo, Lorenzo [2 ]
Madani, Mohammad H. [2 ]
机构
[1] Calif Northstate Univ, Coll Med, 9700 W Taron Dr, Elk Grove, CA 95757 USA
[2] Univ Calif Davis, Med Ctr, Dept Radiol, 4860 Y St,Suite 3100, Sacramento, CA 95817 USA
[3] Univ Calif Davis, Med Ctr, Div Cardiovasc Med, 4860 Y St,Suite 0200, Sacramento, CA 95817 USA
关键词
Accuracy; Cardiac imaging; ChatGPT; Patient education; EXPERT CONSENSUS DOCUMENT; COMPUTED-TOMOGRAPHY SCCT; CORONARY-ARTERY-DISEASE; AMERICAN-COLLEGE; RADIOLOGY ACR; SOCIETY;
D O I
10.1016/j.clinimag.2024.110193
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Purpose: To assess ChatGPT 's ability as a resource for educating patients on various aspects of cardiac imaging, including diagnosis, imaging modalities, indications, interpretation of radiology reports, and management. Methods: 30 questions were posed to ChatGPT-3.5 and ChatGPT-4 three times in three separate chat sessions. Responses were scored as correct, incorrect, or clinically misleading categories by three observers -two board certified cardiologists and one board certified radiologist with cardiac imaging subspecialization. Consistency of responses across the three sessions was also evaluated. Final categorization was based on majority vote between at least two of the three observers. Results: ChatGPT-3.5 answered seventeen of twenty eight questions correctly (61 %) by majority vote. Twenty one of twenty eight questions were answered correctly (75 %) by ChatGPT-4 by majority vote. Majority vote for correctness was not achieved for two questions. Twenty six of thirty questions were answered consistently by ChatGPT-3.5 (87 %). Twenty nine of thirty questions were answered consistently by ChatGPT-4 (97 %). ChatGPT-3.5 had both consistent and correct responses to seventeen of twenty eight questions (61 %). ChatGPT-4 had both consistent and correct responses to twenty of twenty eight questions (71 %). Conclusion: ChatGPT-4 had overall better performance than ChatGTP-3.5 when answering cardiac imaging questions with regard to correctness and consistency of responses. While both ChatGPT-3.5 and ChatGPT-4 answers over half of cardiac imaging questions correctly, inaccurate, clinically misleading and inconsistent responses suggest the need for further refinement before its application for educating patients about cardiac imaging.
引用
收藏
页数:8
相关论文
共 29 条
  • [1] Cardiovascular Toxicity Related to Cancer Treatment: A Pragmatic Approach to the American and European Cardio-Oncology Guidelines
    Alexandre, Joachim
    Cautela, Jennifer
    Ederhy, Stephane
    Damaj, Ghandi Laurent
    Salem, Joe-Elie
    Barlesi, Fabrice
    Farnault, Laure
    Charbonnier, Aude
    Mirabel, Mariana
    Champiat, Stephane
    Cohen-Solal, Alain
    Cohen, Ariel
    Dolladille, Charles
    Thuny, Franck
    [J]. JOURNAL OF THE AMERICAN HEART ASSOCIATION, 2020, 9 (18):
  • [2] Clinical Impact of Residual Leaks Following Left Atrial Appendage Occlusion Insights From the NCDR LAAO Registry
    Alkhouli, Mohamad
    Du, Chengan
    Killu, Ammar
    Simard, Trevor
    Noseworthy, Peter A.
    Friedman, Paul A.
    Curtis, Jeptha P.
    V. Freeman, James
    Holmes, David R.
    [J]. JACC-CLINICAL ELECTROPHYSIOLOGY, 2022, 8 (06) : 766 - 778
  • [3] Amisha, 2019, J FAM MED PRIM CARE, V8, P2328, DOI DOI 10.4103/jfmpc.jfmpc_440_19
  • [4] [Anonymous], Contrast Manual Internet
  • [5] [Anonymous], Introducing ChatGPT Internet
  • [6] Challenges to the Reproducibility of Machine Learning Models in Health Care
    Beam, Andrew L.
    Manrai, Arjun K.
    Ghassemi, Marzyeh
    [J]. JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 2020, 323 (04): : 305 - 306
  • [7] ACC/AHA/ASE/ASNC/ASPC/HFSA/ HRS/SCAI/SCCT/SCMR/STS 2023 Multimodality Appropriate Use Criteria for the Detection and Risk Assessment of Chronic Coronary Disease
    Bhave, Nicole
    [J]. JOURNAL OF THE AMERICAN COLLEGE OF CARDIOLOGY, 2023, 81 (25) : 2445 - 2467
  • [8] Joint SNMMI-ASNC Expert Consensus Document on the Role of 18F-FDG PET/CT in Cardiac Sarcoid Detection and Therapy Monitoring
    Chareonthaitawee, Panithaya
    Beanlands, Rob S.
    Chen, Wengen
    Dorbala, Sharmila
    Miller, Edward J.
    Murthy, Venkatesh L.
    Birnie, David H.
    Chen, Edward S.
    Cooper, Leslie T.
    Tung, Roderick H.
    White, Eric S.
    Borges-Neto, Salvador
    Di Carli, Marcelo F.
    Gropler, Robert J.
    Ruddy, Terrence D.
    Schindler, Thomas H.
    Blankstein, Ron
    [J]. JOURNAL OF NUCLEAR MEDICINE, 2017, 58 (08) : 1341 - 1353
  • [9] Chen S, 2023, medRxiv, DOI [10.1101/2023.03.16.23287316, 10.1101/2023.03.16.23287316v1, DOI 10.1101/2023.03.16.23287316, 10.1101/2023.03.16.23287316]
  • [10] CAD-RADSTM 2.0-2022 Coronary Artery Disease-Reporting and Data System An Expert Consensus Document of the Society of Cardiovascular Computed Tomography (SCCT), the American College of Cardiology (ACC), the American College of Radiology (ACR), and the North America Society of Cardiovascular Imaging (NASCI)
    Cury, Ricardo C.
    Leipsic, Jonathon
    Abbara, Suhny
    Achenbach, Stephan
    Berman, Daniel
    Bittencourt, Marcio
    Budoff, Matthew
    Chinnaiyan, Kavitha
    Choi, Andrew D.
    Ghoshhajra, Brian
    Jacobs, Jill
    Koweek, Lynne
    Lesser, John
    Maroules, Christopher
    Rubin, Geoffrey D.
    Rybicki, Frank J.
    Shaw, Leslee J.
    Williams, Michelle C.
    Williamson, Eric
    White, Charles S.
    Villines, Todd C.
    Blankstein, Ron
    [J]. JOURNAL OF CARDIOVASCULAR COMPUTED TOMOGRAPHY, 2022, 16 (06) : 536 - 557