Accuracy of ChatGPT responses on tracheotomy for patient education

被引:4
作者
Khaldi, Amina [1 ]
Machayekhi, Shahram [1 ]
Salvagno, Michele [2 ]
Maniaci, Antonino [3 ]
Vaira, Luigi A. [4 ]
La Via, Luigi [5 ]
Taccone, Fabio S. [2 ]
Lechien, Jerome R. [6 ,7 ,8 ]
机构
[1] EpiCURA Hosp, Intens Care Unit, Hornu, Belgium
[2] Erasme Univ Hosp, Intens Care Unit, Brussels, Belgium
[3] Univ Enna Kore, Fac Med & Surg, I-94100 Enna, Italy
[4] Univ Sassari, Dept Med Surg & Pharm, Maxillofacial Surg Operat Unit, Viale San Pietro 43-B, I-07100 Sassari, Italy
[5] Univ Hosp Policlin GRodol San Marco, Dept Anesthesia & Intens Care, Catania, Italy
[6] Univ Mons UMons, UMONS Res Inst Hlth Sci & Technol, Dept Surg, Mons, Belgium
[7] Elsan Polyclin Poitiers, Dept Otolaryngol, Poitiers, France
[8] Paris Saclay Univ, Univ Versailles St Quentin en Yvelines, Foch Hosp,UFR Simone Veil, Sch Med,Dept Otolaryngol Head Neck Surg, Paris, France
关键词
ChatGPT; LLM; Artificial intelligence; Tracheotomy; Intensive care; Otolaryngology; Head neck; Surgery; Information; Accuracy;
D O I
10.1007/s00405-024-08859-8
中图分类号
R76 [耳鼻咽喉科学];
学科分类号
100213 ;
摘要
ObjectiveTo investigate the accuracy of information provided by ChatGPT-4o to patients about tracheotomy.MethodsTwenty common questions of patients about tracheotomy were presented to ChatGPT-4o twice (7-day intervals). The accuracy, clarity, relevance, completeness, referencing, and usefulness of responses were assessed by a board-certified otolaryngologist and a board-certified intensive care unit practitioner with the Quality Analysis of Medical Artificial Intelligence (QAMAI) tool. The interrater reliability and the stability of the ChatGPT-4o responses were evaluated with intraclass correlation coefficient (ICC) and Pearson correlation analysis.ResultsThe total scores of QAMAI were 22.85 +/- 4.75 for the intensive care practitioner and 21.45 +/- 3.95 for the otolaryngologist, which consists of moderate-to-high accuracy. The otolaryngologist and the ICU practitioner reported high ICC (0.807; 95%CI: 0.655-0.911). The highest QAMAI scores have been found for clarity and completeness of explanations. The QAMAI scores for the accuracy of the information and the referencing were the lowest. The information related to the post-laryngectomy tracheostomy remains incomplete or erroneous. ChatGPT-4o did not provide references for their responses. The stability analysis reported high stability in regenerated questions.ConclusionThe accuracy of ChatGPT-4o is moderate-to-high in providing information related to the tracheotomy. However, patients using ChatGPT-4o need to be cautious about the information related to tracheotomy care, steps, and the differences between temporary and permanent tracheotomies.
引用
收藏
页码:6167 / 6172
页数:6
相关论文
共 50 条
[31]   Comparison of ChatGPT Models in Patient Education on Obstructive Sleep Apnea [J].
Remzi Doğan ;
Ramazan Bahadır Küçük ;
Orhan Özturan ;
Fadlullah Aksoy ;
Sabri Baki Eren ;
Alper Yenigün ;
Erol Şentürk .
SN Comprehensive Clinical Medicine, 7 (1)
[32]   Accuracy of ChatGPT in Neurolocalization [J].
Dabbas, Waleed F. ;
Odeibat, Yousef M. ;
Alhazaimeh, Mohammad ;
Hiasat, Mohammad Y. ;
Alomari, Amer A. ;
Marji, Ala ;
Samara, Qais A. ;
Ibrahim, Bilal ;
Al Arabiyat, Rashed M. ;
Momani, Ghena .
CUREUS JOURNAL OF MEDICAL SCIENCE, 2024, 16 (04)
[33]   Assessing the Accuracy of Responses by the Language Model ChatGPT to Questions Regarding Bariatric Surgery [J].
Jamil S. Samaan ;
Yee Hui Yeo ;
Nithya Rajeev ;
Lauren Hawley ;
Stuart Abel ;
Wee Han Ng ;
Nitin Srinivasan ;
Justin Park ;
Miguel Burch ;
Rabindra Watson ;
Omer Liran ;
Kamran Samakar .
Obesity Surgery, 2023, 33 :1790-1796
[34]   ChatGPT: the essential patient's companion for transesophageal echocardiogram education [J].
Avidan, Yuval ;
Sliman, Hussein ;
Weizman, Baruch ;
Ben Court, Orel ;
Fuks, Alexander ;
Adawi, Salim ;
Aker, Amir .
JOURNAL OF ULTRASOUND, 2025,
[35]   Assessing the accuracy and reliability of ChatGPT's medical responses about thyroid cancer [J].
Helvaci, Burcak Cavnar ;
Hepsen, Sema ;
Candemir, Burcu ;
Boz, Ogulcan ;
Durantas, Halil ;
Houssein, Mehdi ;
Cakal, Erman .
INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 2024, 191
[36]   Comment on "Evaluating ChatGPT's Accuracy in Responding to Patient Education Questions on Acute Kidney Injury and Continuous Renal Replacement Therapy" [J].
Daungsupawong, Hinpetch ;
Wiwanitkit, Viroj .
BLOOD PURIFICATION, 2024, 53 (10) :847-848
[37]   Comparative evaluation of ChatGPT and LLaMA for reliability, quality, and accuracy in familial Mediterranean fever [J].
Bektas, Aslihan Uzun ;
Bora, Balahan ;
Unsal, Erbil .
EUROPEAN JOURNAL OF PEDIATRICS, 2025, 184 (08)
[38]   Evaluating the accuracy of ChatGPT in delivering patient instructions for medications: an exploratory case study [J].
Abanmy, Norah Othman ;
Al-Ghreimil, Nadia ;
Alsabhan, Jawza F. ;
Al-Baity, Heyam ;
Aljadeed, Rana .
FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2025, 8
[39]   The utility of ChatGPT in generating patient-facing and clinical responses for melanoma [J].
Young, Jade N. ;
O'Hagan, Ross ;
Poplausky, Dina ;
Levoska, Melissa A. ;
Gulati, Nicholas ;
Ungar, Benjamin ;
Ungar, Jonathan .
JOURNAL OF THE AMERICAN ACADEMY OF DERMATOLOGY, 2023, 89 (03) :602-604
[40]   Evaluation of Google and ChatGPT responses to common patient questions about scoliosis [J].
Tekin, Sezgin Bahadir ;
Ince, Kamil ;
Tekin, Bedriye Gizem ;
Servet, Erkan ;
Bozgeyik, Bahri .
SPINE DEFORMITY, 2025,