Accuracy of ChatGPT responses on tracheotomy for patient education

被引:2
|
作者
Khaldi, Amina [1 ]
Machayekhi, Shahram [1 ]
Salvagno, Michele [2 ]
Maniaci, Antonino [3 ]
Vaira, Luigi A. [4 ]
La Via, Luigi [5 ]
Taccone, Fabio S. [2 ]
Lechien, Jerome R. [6 ,7 ,8 ]
机构
[1] EpiCURA Hosp, Intens Care Unit, Hornu, Belgium
[2] Erasme Univ Hosp, Intens Care Unit, Brussels, Belgium
[3] Univ Enna Kore, Fac Med & Surg, I-94100 Enna, Italy
[4] Univ Sassari, Dept Med Surg & Pharm, Maxillofacial Surg Operat Unit, Viale San Pietro 43-B, I-07100 Sassari, Italy
[5] Univ Hosp Policlin GRodol San Marco, Dept Anesthesia & Intens Care, Catania, Italy
[6] Univ Mons UMons, UMONS Res Inst Hlth Sci & Technol, Dept Surg, Mons, Belgium
[7] Elsan Polyclin Poitiers, Dept Otolaryngol, Poitiers, France
[8] Paris Saclay Univ, Univ Versailles St Quentin en Yvelines, Foch Hosp,UFR Simone Veil, Sch Med,Dept Otolaryngol Head Neck Surg, Paris, France
关键词
ChatGPT; LLM; Artificial intelligence; Tracheotomy; Intensive care; Otolaryngology; Head neck; Surgery; Information; Accuracy;
D O I
10.1007/s00405-024-08859-8
中图分类号
R76 [耳鼻咽喉科学];
学科分类号
100213 ;
摘要
ObjectiveTo investigate the accuracy of information provided by ChatGPT-4o to patients about tracheotomy.MethodsTwenty common questions of patients about tracheotomy were presented to ChatGPT-4o twice (7-day intervals). The accuracy, clarity, relevance, completeness, referencing, and usefulness of responses were assessed by a board-certified otolaryngologist and a board-certified intensive care unit practitioner with the Quality Analysis of Medical Artificial Intelligence (QAMAI) tool. The interrater reliability and the stability of the ChatGPT-4o responses were evaluated with intraclass correlation coefficient (ICC) and Pearson correlation analysis.ResultsThe total scores of QAMAI were 22.85 +/- 4.75 for the intensive care practitioner and 21.45 +/- 3.95 for the otolaryngologist, which consists of moderate-to-high accuracy. The otolaryngologist and the ICU practitioner reported high ICC (0.807; 95%CI: 0.655-0.911). The highest QAMAI scores have been found for clarity and completeness of explanations. The QAMAI scores for the accuracy of the information and the referencing were the lowest. The information related to the post-laryngectomy tracheostomy remains incomplete or erroneous. ChatGPT-4o did not provide references for their responses. The stability analysis reported high stability in regenerated questions.ConclusionThe accuracy of ChatGPT-4o is moderate-to-high in providing information related to the tracheotomy. However, patients using ChatGPT-4o need to be cautious about the information related to tracheotomy care, steps, and the differences between temporary and permanent tracheotomies.
引用
收藏
页码:6167 / 6172
页数:6
相关论文
共 50 条
  • [1] Readability, accuracy, and appropriateness of ChatGPT 4.0 responses for use in patient education materials for Condyloma acuminatum
    Moosvi, Nosheen
    Kovarik, Carrie
    CLINICS IN DERMATOLOGY, 2024, 42 (01) : 87 - 88
  • [2] Evaluating Chatgpt Responses on Atrial Fibrillation for Patient Education
    Lee, Thomas J.
    Campbell, Daniel J.
    Elkattawy, Omar
    Viswanathan, Rohan
    CIRCULATION, 2023, 148
  • [3] Evaluating ChatGPT Responses on Atrial Fibrillation for Patient Education
    Lee, Thomas J.
    Campbell, Daniel J.
    Rao, Abhinav K.
    Hossain, Afif
    Elkattawy, Omar
    Radfar, Navid
    Lee, Paul
    Gardin, Julius M.
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2024, 16 (06)
  • [4] Evaluating ChatGPT Responses on Thyroid Nodules for Patient Education
    Campbell, Daniel J.
    Estephan, Leonard E.
    Sina, Elliott M.
    Mastrolonardo, Eric V.
    Alapati, Rahul
    Amin, Dev R.
    Cottrill, Elizabeth E.
    THYROID, 2024, 34 (03) : 371 - 377
  • [5] ChatGPT-4 accuracy for patient education in laryngopharyngeal reflux
    Lechien, Jerome R.
    Carroll, Thomas L.
    Huston, Molly N.
    Naunheim, Matthew R.
    EUROPEAN ARCHIVES OF OTO-RHINO-LARYNGOLOGY, 2024, 281 (05) : 2547 - 2552
  • [6] ChatGPT-4 accuracy for patient education in laryngopharyngeal reflux
    Jerome R. Lechien
    Thomas L. Carroll
    Molly N. Huston
    Matthew R. Naunheim
    European Archives of Oto-Rhino-Laryngology, 2024, 281 : 2547 - 2552
  • [7] Evaluating ChatGPT-3.5 and ChatGPT-4.0 Responses on Hyperlipidemia for Patient Education
    Lee, Thomas J.
    Rao, Abhinav K.
    Campbell, Daniel J.
    Radfar, Navid
    Dayal, Manik
    Khrais, Ayham
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2024, 16 (05)
  • [8] Accuracy and Readability of ChatGPT Responses to Patient-Centric Strabismus Questions
    Gary, Ashlyn A.
    Lai, James M.
    Locatelli, Elyana V. T.
    Falcone, Michelle M.
    Cavuoto, Kara M.
    JOURNAL OF PEDIATRIC OPHTHALMOLOGY & STRABISMUS, 2025,
  • [9] Evaluating ChatGPT responses on obstructive sleep apnea for patient education
    Campbell, Daniel J.
    Estephan, Leonard E.
    Mastrolonardo, Eric V.
    Amin, Dev R.
    Huntley, Colin T.
    Boon, Maurits S.
    JOURNAL OF CLINICAL SLEEP MEDICINE, 2023, 19 (12): : 1989 - 1995
  • [10] COMPARING RESPONSES FROM CHATGPT AND BARD ON AORTIC STENOSIS FOR PATIENT EDUCATION
    Cotugno, Lorenzo R.
    Lee, Thomas
    Patel, Sarthak
    JOURNAL OF THE AMERICAN COLLEGE OF CARDIOLOGY, 2024, 83 (13) : 2682 - 2682