Improving Patient Understanding of Glomerular Disease Terms With ChatGPT

被引:0
作者
Abdelgadir, Yasir H. [1 ]
Thongprayoon, Charat [1 ]
Craici, Iasmina M. [1 ]
Cheungpasitporn, Wisit [1 ]
Miao, Jing [1 ]
机构
[1] Mayo Clin, Dept Med, Div Nephrol & Hypertens, Rochester, MN 55905 USA
关键词
accuracy; ChatGPT; glomerular disease; interpretation; patient education; readability; READABILITY;
D O I
10.1155/ijcp/9977290
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Background: Glomerular disease is complex and difficult for patients to understand, as it involves various pathophysiology, immunology, and pharmacology areas. Objective: This study explored whether ChatGPT can maintain accuracy while simplifying glomerular disease terms to enhance patient comprehension. Methods: 67 terms related to glomerular disease were analyzed using GPT-4 through two distinct queries. One aimed at a general explanation and another tailored for patients with an education level of 8th grade or lower. GPT-4's accuracy was scored from 1 (incorrect) to 5 (correct and comprehensive). Its readability was assessed using the Consensus Reading Grade (CRG) Level, which incorporates seven readability indices including the Flesch-Kincaid Grade (FKG) and SMOG indices. Flesch Reading Ease (FRE) score, ranging from 0 to 100 with higher scores indicating easier-to-read text, was also used to evaluate the readability. A paired t-test was conducted to assess differences in accuracy and readability levels between different queries. Results: GPT-4's general explanations of glomerular disease terms averaged at a college readability level, indicated by the CRG score of 14.1 and FKG score of 13.9. SMOG index also indicated the topic's complexity, with a score of 11.8. When tailored for patients at or below an 8(th)-grade reading level, readability improved, averaging 9.7 by the CRG score, 8.7 by FKG score, and 7.3 by SMOG score. The FRE score also indicated a further improvement of readability from 31.6 for general explanations to 63.5 for tailored explanations. However, the accuracy in GPT-4's tailored explanations was significantly lower than that in general explanations (4.2 +/- 0.4 versus 4.7 +/- 0.3, p < 0.0001). Conclusion: While GPT-4 effectively simplified information about glomerular diseases, it compromised its accuracy in the process. To implement these findings, we suggest pilot studies in clinical settings to assess patient understanding, using feedback from diverse groups to customize content, expanding research to enhance AI accuracy and reduce biases, setting strict ethical guidelines for AI in healthcare, and integrating with health informatics systems to provide tailored educational content to patients. This approach will promote effective and ethical use of AI tools like ChatGPT in patient education, empowering patients to make informed health decisions.
引用
收藏
页数:7
相关论文
共 25 条
[1]   Enhancing Readability of Online Patient-Facing Content: The Role of AI Chatbots in Improving Cancer Information Accessibility [J].
Abreu, Andres A. ;
Murimwa, Gilbert Z. ;
Farah, Emile ;
Stewart, James W. ;
Zhang, Lucia ;
Rodriguez, Jonathan ;
Sweetenham, John ;
Zeh, Herbert J. ;
Wang, Sam C. ;
Polanco, Patricio M. .
JOURNAL OF THE NATIONAL COMPREHENSIVE CANCER NETWORK, 2024, 22 (2D)
[2]   Limitations of readability assessment tools [J].
Alzaid, Mohammad ;
Ali, Faisal R. ;
Stapleton, Emma .
EUROPEAN ARCHIVES OF OTO-RHINO-LARYNGOLOGY, 2024, 281 (09) :5021-5022
[3]   Large language models in patient education: a scoping review of applications in medicine [J].
Aydin, Serhat ;
Karabacak, Mert ;
Vlachos, Victoria ;
Margetis, Konstantinos .
FRONTIERS IN MEDICINE, 2024, 11
[4]  
Browne Robert, 2024, J Hand Surg Glob Online, V6, P436, DOI 10.1016/j.jhsg.2024.03.008
[5]   Literacy and health outcomes - A systematic review of the literature [J].
DeWalt, DA ;
Berkman, ND ;
Sheridan, S ;
Lohr, KN ;
Pignone, MP .
JOURNAL OF GENERAL INTERNAL MEDICINE, 2004, 19 (12) :1228-1239
[6]   Improving Readability and Automating Content Analysis of Plastic Surgery Webpages With ChatGPT [J].
Fanning, James E. ;
Escobar-Domingo, Maria J. ;
Foppiani, Jose ;
Lee, Daniela ;
Miller, Amitai S. ;
Janis, Jeffrey E. ;
Lee, Bernard T. .
JOURNAL OF SURGICAL RESEARCH, 2024, 299 :103-111
[7]  
Feehally J., 2019, Comprehensive Clinical Nephrology, P184
[8]   Readability assessment of concussion and traumatic brain injury publications by Centers for Disease Control and Prevention [J].
Gill, Preetinder S. ;
Gill, Tejkaran S. ;
Kamath, Ashwini ;
Whisnant, Billy .
INTERNATIONAL JOURNAL OF GENERAL MEDICINE, 2012, 5 :923-933
[9]  
Graphpad, 2024, GraphPad
[10]   Accuracy, readability, and understandability of large language models for prostate cancer information to the public [J].
Hershenhouse, Jacob S. ;
Mokhtar, Daniel ;
Eppler, Michael B. ;
Rodler, Severin ;
Ramacciotti, Lorenzo Storino ;
Ganjavi, Conner ;
Hom, Brian ;
Davis, Ryan J. ;
Tran, John ;
Russo, Giorgio Ivan ;
Cocci, Andrea ;
Abreu, Andre ;
Gill, Inderbir ;
Desai, Mihir ;
Cacciamani, Giovanni E. .
PROSTATE CANCER AND PROSTATIC DISEASES, 2024, :394-399