Comparative Analysis of Large Language Models in Simplifying Turkish Ultrasound Reports to Enhance Patient Understanding

被引:2
作者
Gunes, Yasin Celal [1 ]
Cesur, Turay [2 ]
Camur, Eren [3 ]
机构
[1] Kirikkale Yuksek Ihtisas Hosp, Dept Radiol, Ahmet Ay Caddesi, TR-71300 Kirikkale, Turkiye
[2] Ankara Mamak State Hosp, Dept Radiol, Ankara, Turkiye
[3] Ankara 29 Mayis State Hosp, Dept Radiol, Ankara, Turkiye
来源
EUROPEAN JOURNAL OF THERAPEUTICS | 2024年 / 30卷 / 05期
关键词
Large Language Models; ChatGPT; Claude; 3; Opus; Ultrasound; Simplify; RADIOLOGY REPORTS; CHATGPT;
D O I
10.58600/eurjther2225
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Objective: To evaluate and compare the abilities of Language Models (LLMs) in simplifying Turkish ultrasound (US) findings for patients. Methods: We assessed the simplification performance of four LLMs: ChatGPT 4, Gemini 1.5 Pro, Claude 3 Opus, and Perplexity, using fifty fictional Turkish US findings. Comparison was based on Ate man's Readability Index and word count. Three radiologists rated medical accuracy, consistency, and comprehensibility on a Likert scale from 1 to 5. Statistical tests (Friedman, Wilcoxon, and Spearman correlation) examined differences in LLMs' performance. Results: Gemini 1.5 Pro, ChatGPT-4, and Claude 3 Opus received high Likert scores for medical accuracy, consistency, and comprehensibility (mean: 4.7-4.8). Perplexity scored significantly lower (mean: 4.1, p<0.001). Gemini 1.5 Pro achieved the highest readability score (mean: 61.16), followed by ChatGPT-4 (mean: 58.94) and Claude 3 Opus (mean: 51.16). Perplexity had the lowest readability score (mean: 47.01). Gemini 1.5 Pro and ChatGPT-4 used significantly more words compared to Claude 3 Opus and Perplexity (p<0.001). Linear correlation analysis revealed a positive correlation between word count of fictional US findings and responses generated by Gemini 1.5 Pro (correlation coefficient = 0.38, p<0.05) and ChatGPT-4 (correlation coefficient = 0.43, p<0.001). Conclusion: This study highlights strong potential of LLMs in simplifying Turkish US and Claude 3 Opus performed well, highlighting their effectiveness in healthcare communication. Further research is required to fully understand the integration of making.
引用
收藏
页码:714 / 723
页数:10
相关论文
共 21 条
[1]   Potential Use Cases for ChatGPT in Radiology Reporting [J].
Abou Elkassem, Asser ;
Smith, Andrew D. .
AMERICAN JOURNAL OF ROENTGENOLOGY, 2023, 221 (03) :373-376
[2]  
Amin K, 2023, YALE J BIOL MED, V96, P407, DOI 10.59249/NKOY5498
[3]  
Atesman E, 1997, DERGISI, V58
[4]  
Aydin Omer, 2023, Acad. Platf. J. Eng. Smart Syst, V11, P118, DOI [10.21541/apjess.1293702, DOI 10.21541/APJESS.1293702]
[5]  
Bossuyt PM, 2015, BMJ-BRIT MED J, V351, DOI [10.1148/radiol.2015151516, 10.1136/bmj.h5527, 10.1373/clinchem.2015.246280]
[6]  
Chan V, 2011, ATLAS OF ULTRASOUND-GUIDED PROCEDURES IN INTERVENTIONAL PAIN MANAGEMENT, P13, DOI 10.1007/978-1-4419-1681-5_2
[7]   Feasibility and acceptability of ChatGPT generated radiology report summaries for cancer patients [J].
Chung, Eric M. ;
Zhang, Samuel C. ;
Nguyen, Anthony T. ;
Atkins, Katelyn M. ;
Sandler, Howard M. ;
Kamrava, Mitchell .
DIGITAL HEALTH, 2023, 9
[8]   Quantitative Evaluation of Large Language Models to Streamline Radiology Report Impressions: A Multimodal Retrospective Analysis [J].
Doshi, Rushabh ;
Amin, Kanhai S. ;
Khosla, Pavan ;
Bajaj, Simar S. ;
Chheang, Sophie ;
Forman, Howard P. .
RADIOLOGY, 2024, 310 (03)
[9]   Challenging ChatGPT 3.5 in Senology-An Assessment of Concordance with Breast Cancer Tumor Board Decision Making [J].
Griewing, Sebastian ;
Gremke, Niklas ;
Wagner, Uwe ;
Lingenfelder, Michael ;
Kuhn, Sebastian ;
Boekhoff, Jelena .
JOURNAL OF PERSONALIZED MEDICINE, 2023, 13 (10)
[10]   Evaluating the Use of ChatGPT to Accurately Simplify Patient-centered Information about Breast Cancer Prevention and Screening [J].
Haver, Hana L. ;
Gupta, Anuj K. ;
Ambinder, Emily B. ;
Bahl, Manisha ;
Oluyemi, Eniola T. ;
Jeudy, Jean ;
Yi, Paul H. .
RADIOLOGY-IMAGING CANCER, 2024, 6 (02)