Enhancing Readability of Online Patient-Facing Content: The Role of AI Chatbots in Improving Cancer Information Accessibility

被引:12
作者
Abreu, Andres A. [1 ]
Murimwa, Gilbert Z. [1 ]
Farah, Emile [1 ]
Stewart, James W. [2 ]
Zhang, Lucia [1 ]
Rodriguez, Jonathan [1 ]
Sweetenham, John [1 ]
Zeh, Herbert J. [1 ]
Wang, Sam C. [1 ]
Polanco, Patricio M. [1 ]
机构
[1] UT Southwestern Med Ctr, Dept Surg, Div Surg Oncol, 5323 Harry Hines Blvd, Dallas, TX 75390 USA
[2] Yale Sch Med, Dept Surg, New Haven, CT USA
来源
JOURNAL OF THE NATIONAL COMPREHENSIVE CANCER NETWORK | 2024年 / 22卷 / 2D期
关键词
HEALTH LITERACY; ASSOCIATION; INTERNET; WEB;
D O I
10.6004/jnccn.2023.7334
中图分类号
R73 [肿瘤学];
学科分类号
100214 ;
摘要
Background: Internet-based health education is increasingly vital in patient care. However, the readability of online information often exceeds the average reading level of the US population, limiting accessibility and comprehension. This study investigates the use of chatbot artificial intelligence to improve the readability of cancer-related patient-facing content. Methods: We used ChatGPT 4.0 to rewrite content about breast, colon, lung, prostate, and pancreas cancer across 34 websites associated with NCCN Member Institutions. Readability was analyzed using Fry Readability Score, Flesch-Kincaid Grade Level, Gunning Fog Index, and Simple Measure of Gobbledygook. The primary outcome was the mean readability score for the original and artificial intelligence (AI)-generated content. As secondary outcomes, we assessed the accuracy, similarity, and quality using F1 scores, cosine similarity scores, and section 2 of the DISCERN instrument, respectively. Results: The mean readability level across the 34 websites was equivalent to a university freshman level (grade 13 +/- 1.5). However, after ChatGPT's intervention, the AI-generated outputs had a mean readability score equivalent to a high school freshman education level (grade 9 +/- 0.8). The overall F1 score for the rewritten content was 0.87, the precision score was 0.934, and the recall score was 0.814. Compared with their original counterparts, the AI-rewritten content had a cosine similarity score of 0.915 (95% CI, 0.908-0.922). The improved readability was attributed to simpler words and shorter sentences. The mean DISCERN score of the random sample of AI-generated content was equivalent to "good" (28.5 +/- 5), with no significant differences compared with their original counterparts. Conclusions: Our study demonstrates the potential of AI chatbots to improve the readability of patient-facing content while maintaining content quality. The decrease in requisite literacy after AI revision emphasizes the potential of this technology to reduce health care disparities caused by a mismatch between educational resources available to a patient and their health literacy.
引用
收藏
页数:8
相关论文
共 47 条
[1]   Comparing Physician and Artificial Intelligence Chatbot Responses to Patient Questions Posted to a Public Social Media Forum [J].
Ayers, John W. ;
Poliak, Adam ;
Dredze, Mark ;
Leas, Eric C. ;
Zhu, Zechariah ;
Kelley, Jessica B. ;
Faix, Dennis J. ;
Goodman, Aaron M. ;
Longhurst, Christopher A. ;
Hogarth, Michael ;
Smith, Davey M. .
JAMA INTERNAL MEDICINE, 2023, 183 (06) :589-596
[2]   Use of the Internet and e-mail for health care information - Results from a national survey [J].
Baker, L ;
Wagner, TH ;
Singer, S ;
Bundorf, MK .
JAMA-JOURNAL OF THE AMERICAN MEDICAL ASSOCIATION, 2003, 289 (18) :2400-2406
[3]   How Good Is Online Information for Patients on the Treatment for Luminal Gastrointestinal Cancers? A Comprehensive Evaluation in English and Spanish [J].
Beas, Renato ;
Cabanillas-Ramirez, Cielo ;
Izquierdo-Veraza, Diego ;
Chaponan-Lavalle, Andres ;
Norwood, Dalton ;
Sato-Espinoza, Karina ;
Riva-Moscoso, Adrian ;
Ribaudo, Isabella ;
Ramirez-Rojas, Mirian ;
Beran, Azizullah ;
Montalvan-Sanchez, Eleazar E. .
JOURNAL OF CANCER EDUCATION, 2023, 38 (06) :1801-1807
[4]  
Brach C, AHRQ health literacy universal precautions toolkit, V3rd
[5]   Artificial intelligence and urology: ethical considerations for urologists and patients [J].
Cacciamani, Giovanni E. ;
Chen, Andrew ;
Gill, Inderbir S. ;
Hung, Andrew J. J. .
NATURE REVIEWS UROLOGY, 2024, 21 (01) :50-59
[6]   DISCERN: an instrument for judging the quality of written consumer health information on treatment choices [J].
Charnock, D ;
Shepperd, S ;
Needham, G ;
Gann, R .
JOURNAL OF EPIDEMIOLOGY AND COMMUNITY HEALTH, 1999, 53 (02) :105-111
[7]   Use of Artificial Intelligence Chatbots for Cancer Treatment Information [J].
Chen, Shan ;
Kann, Benjamin H. ;
Foote, Michael B. ;
Aerts, Hugo J. W. L. ;
Savova, Guergana K. ;
Mak, Raymond H. ;
Bitterman, Danielle S. .
JAMA ONCOLOGY, 2023, 9 (10) :1459-1462
[8]   Quality of information and appropriateness of ChatGPT outputs for urology patients [J].
Cocci, Andrea ;
Pezzoli, Marta ;
Lo Re, Mattia ;
Russo, Giorgio Ivan ;
Asmundo, Maria Giovanna ;
Fode, Mikkel ;
Cacciamani, Giovanni ;
Cimino, Sebastiano ;
Minervini, Andrea ;
Durukan, Emil .
PROSTATE CANCER AND PROSTATIC DISEASES, 2024, 27 (01) :103-108
[9]  
Cohen R A., 2009, Health Information Technology Use among Men and Women aged 18-64: Early Release of Estimates from the National Health Interview Survey, January-June 2009
[10]   Can ChatGPT, an Artificial Intelligence Language Model, Provide Accurate and High-quality Patient Information on Prostate Cancer? [J].
Coskun, Burhan ;
Ocakoglu, Gokhan ;
Yetemen, Melih ;
Kaygisiz, Onur .
UROLOGY, 2023, 180 :35-58