Artificial intelligence in global health equity: an evaluation and discussion on the application of ChatGPT, in the Chinese National Medical Licensing Examination

被引:8
作者
Tong, Wenting [1 ]
Guan, Yongfu [2 ]
Chen, Jinping [2 ]
Huang, Xixuan [3 ]
Zhong, Yuting [4 ]
Zhang, Changrong [5 ]
Zhang, Hui [2 ,6 ]
机构
[1] Gannan Healthcare Vocat Coll, Dept Pharm, Ganzhou, Peoples R China
[2] Gannan Hlth Vocat Coll, Dept Rehabil & Elderly Care, Ganzhou, Jiangxi, Peoples R China
[3] Xiamen Univ, Dept Math, Xiamen, Fujian, Peoples R China
[4] Gannan Med Univ, Dept Anesthesiol, Ganzhou, Jiangxi, Peoples R China
[5] Qinghai Univ, Dept Chinese Med, Affiliated Hosp, Xining, Qinghai, Peoples R China
[6] Univ Roma Tor Vergata, Dept Syst Med, Chair Endocrinol & Med Sexol ENDOSEX, Rome, Italy
关键词
global healthcare; equity; artificial intelligence; ChatGPT; language bias;
D O I
10.3389/fmed.2023.1237432
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
BackgroundThe demand for healthcare is increasing globally, with notable disparities in access to resources, especially in Asia, Africa, and Latin America. The rapid development of Artificial Intelligence (AI) technologies, such as OpenAI's ChatGPT, has shown promise in revolutionizing healthcare. However, potential challenges, including the need for specialized medical training, privacy concerns, and language bias, require attention.MethodsTo assess the applicability and limitations of ChatGPT in Chinese and English settings, we designed an experiment evaluating its performance in the 2022 National Medical Licensing Examination (NMLE) in China. For a standardized evaluation, we used the comprehensive written part of the NMLE, translated into English by a bilingual expert. All questions were input into ChatGPT, which provided answers and reasons for choosing them. Responses were evaluated for "information quality" using the Likert scale.ResultsChatGPT demonstrated a correct response rate of 81.25% for Chinese and 86.25% for English questions. Logistic regression analysis showed that neither the difficulty nor the subject matter of the questions was a significant factor in AI errors. The Brier Scores, indicating predictive accuracy, were 0.19 for Chinese and 0.14 for English, indicating good predictive performance. The average quality score for English responses was excellent (4.43 point), slightly higher than for Chinese (4.34 point).ConclusionWhile AI language models like ChatGPT show promise for global healthcare, language bias is a key challenge. Ensuring that such technologies are robustly trained and sensitive to multiple languages and cultures is vital. Further research into AI's role in healthcare, particularly in areas with limited resources, is warranted.
引用
收藏
页数:7
相关论文
共 22 条
  • [21] Development of an artificial intelligence-based application for the diagnosis of sarcopenia: a retrospective cohort study using the health examination dataset
    Jeong, Chang-Won
    Lim, Dong-Wook
    Noh, Si-Hyeong
    Lee, Sung Hyun
    Park, Chul
    BMC MEDICAL INFORMATICS AND DECISION MAKING, 2025, 25 (01)
  • [22] Exploring the potential of artificial intelligence to enhance the writing of english academic papers by non-native english-speaking medical students - the educational application of ChatGPT
    Li, Jiakun
    Zong, Hui
    Wu, Erman
    Wu, Rongrong
    Peng, Zhufeng
    Zhao, Jing
    Yang, Lu
    Xie, Hong
    Shen, Bairong
    BMC MEDICAL EDUCATION, 2024, 24 (01)