Using Large Language Models to Generate Educational Materials on Childhood Glaucoma

被引:3
作者
Dihan, Qais [1 ,2 ]
Chauhan, Muhammad z. [2 ]
Eleiwa, Taher k. [3 ]
Hassan, Amr k. [4 ]
Sallam, Ahmed b. [2 ,5 ]
Khouri, Albert s. [6 ]
Chang, Ta c. [7 ]
Elhusseiny, Abdelrahman m. [2 ,8 ]
机构
[1] Chicago Med Sch, Dept Med, N Chicago, IL USA
[2] Univ Arkansas Med Sci, Harvey & Bernice Jones Eye Inst, Dept Ophthalmol, Little Rock, AR USA
[3] Univ Arkansas Med Sci, Harvey & Bernice Jones Eye Inst, Benha, AR USA
[4] South Valley Univ, Fac Med, Dept Ophthalmol, Qena, Egypt
[5] Ain Shams Univ, Fac Med, Dept Ophthalmol, Cairo, Egypt
[6] Rutgers New Jersey Med Sch, Inst Ophthalmol & Visual Sci ASK, Newark, NJ USA
[7] Univ Miami, Bascom Palmer Eye Inst, Dept Ophthalmol, Miller Sch Med, Miami, FL USA
[8] Harvard Med Sch, Boston Childrens Hosp, Dept Ophthalmol, Boston, MA USA
关键词
FOLLOW-UP; READABILITY; INFORMATION; ADHERENCE; BARRIERS; QUALITY; CARE;
D O I
10.1016/j.ajo.2024.04.004
中图分类号
R77 [眼科学];
学科分类号
100212 ;
摘要
Purpose: To evaluate the quality, readability, and accuracy of large language model (LLM)-generated patient education materials (PEMs) on childhood glaucoma, and their ability to improve existing the readability of online information. Design: Cross-sectional comparative study. Methods: We evaluated responses of ChatGPT-3.5, ChatGPT-4, and Bard to 3 separate prompts requesting that they write PEMs on "childhood glaucoma." Prompt A required PEMs be "easily understandable by the average American." Prompt B required that PEMs be written "at a 6th-grade level using Simple Measure of Gobbledygook (SMOG) readability formula." We then compared responses' quality (DISCERN questionnaire, Patient Education Materials Assessment Tool [PEMAT]), readability (SMOG, Flesch-Kincaid Grade Level [FKGL]), and accuracy (Likert Misinformation scale). To assess the improvement of readability for existing online information, Prompt C requested that LLM rewrite 20 resources from a Google search of keyword "childhood glaucoma" to the American Medical Association-recommended "6th-grade level." Rewrites were compared on key metrics such as readability, complex words (>= 3 syllables), and sentence count. Results: All 3 LLMs generated PEMs that were of high quality, understandability, and accuracy (DISCERN >= 4, >= 70% PEMAT understandability, Misinformation score = 1). Prompt B responses were more readable than Prompt A responses for all 3 LLM (P <= .001). ChatGPT-4 generated the most readable PEMs compared to ChatGPT-3.5 and Bard (P <= .001). Although Prompt C responses showed consistent reduction of mean SMOG and FKGL scores, only ChatGPT-4 achieved the specified 6th-grade reading level (4.8 +/- 0.8 and 3.7 +/- 1.9, respectively). Conclusions:<bold> </bold>LLMs can serve as strong supplemental tools in generating high-quality, accurate, and novel PEMs, and improving the readability of existing PEMs on childhood glaucoma.
引用
收藏
页码:28 / 38
页数:11
相关论文
共 45 条
  • [1] [Anonymous], 2020, AHRQThe Patient Education Materials Assessment Tool (PEMAT) and User's Guide: introduction
  • [2] [Anonymous], 2010, Health Literacy Universal Precautions Toolkit
  • [3] Comparing Physician and Artificial Intelligence Chatbot Responses to Patient Questions Posted to a Public Social Media Forum
    Ayers, John W.
    Poliak, Adam
    Dredze, Mark
    Leas, Eric C.
    Zhu, Zechariah
    Kelley, Jessica B.
    Faix, Dennis J.
    Goodman, Aaron M.
    Longhurst, Christopher A.
    Hogarth, Michael
    Smith, Davey M.
    [J]. JAMA INTERNAL MEDICINE, 2023, 183 (06) : 589 - 596
  • [4] Comparing ChatGPT and GPT-4 performance in USMLE soft skill assessments
    Brin, Dana
    Sorin, Vera
    Vaid, Akhil
    Soroush, Ali
    Glicksberg, Benjamin S.
    Charney, Alexander W.
    Nadkarni, Girish
    Klang, Eyal
    [J]. SCIENTIFIC REPORTS, 2023, 13 (01)
  • [5] DISCERN: an instrument for judging the quality of written consumer health information on treatment choices
    Charnock, D
    Shepperd, S
    Needham, G
    Gann, R
    [J]. JOURNAL OF EPIDEMIOLOGY AND COMMUNITY HEALTH, 1999, 53 (02) : 105 - 111
  • [6] Analysis of the Readability and Accountability of Online Patient Education Materials Related to Glaucoma Diagnosis and Treatment
    Cohen, Samuel A.
    Fisher, Ann Caroline
    Pershing, Suzann
    [J]. CLINICAL OPHTHALMOLOGY, 2023, 17 : 779 - 788
  • [7] Assessment of the readability and quality of online patient education materials for the medical treatment of open-angle glaucoma
    Crabtree, Lois
    Lee, Edward
    [J]. BMJ OPEN OPHTHALMOLOGY, 2022, 7 (01):
  • [8] Large Language Model-Based Chatbot vs Surgeon-Generated Informed Consent Documentation for Common Procedures
    Decker, Hannah
    Trang, Karen
    Ramirez, Joel
    Colley, Alexis
    Pierce, Logan
    Coleman, Melissa
    Bongiovanni, Tasce
    Melton, Genevieve B.
    Wick, Elizabeth
    [J]. JAMA NETWORK OPEN, 2023, 6 (10)
  • [9] Desktop Search Engine Market Share Worldwide, 2024, StatCounter Global Stats
  • [10] Edwards B, 2023, ARS TECHNICA