Artificial intelligence in global health equity: an evaluation and discussion on the application of ChatGPT, in the Chinese National Medical Licensing Examination

被引:8
作者
Tong, Wenting [1 ]
Guan, Yongfu [2 ]
Chen, Jinping [2 ]
Huang, Xixuan [3 ]
Zhong, Yuting [4 ]
Zhang, Changrong [5 ]
Zhang, Hui [2 ,6 ]
机构
[1] Gannan Healthcare Vocat Coll, Dept Pharm, Ganzhou, Peoples R China
[2] Gannan Hlth Vocat Coll, Dept Rehabil & Elderly Care, Ganzhou, Jiangxi, Peoples R China
[3] Xiamen Univ, Dept Math, Xiamen, Fujian, Peoples R China
[4] Gannan Med Univ, Dept Anesthesiol, Ganzhou, Jiangxi, Peoples R China
[5] Qinghai Univ, Dept Chinese Med, Affiliated Hosp, Xining, Qinghai, Peoples R China
[6] Univ Roma Tor Vergata, Dept Syst Med, Chair Endocrinol & Med Sexol ENDOSEX, Rome, Italy
关键词
global healthcare; equity; artificial intelligence; ChatGPT; language bias;
D O I
10.3389/fmed.2023.1237432
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
BackgroundThe demand for healthcare is increasing globally, with notable disparities in access to resources, especially in Asia, Africa, and Latin America. The rapid development of Artificial Intelligence (AI) technologies, such as OpenAI's ChatGPT, has shown promise in revolutionizing healthcare. However, potential challenges, including the need for specialized medical training, privacy concerns, and language bias, require attention.MethodsTo assess the applicability and limitations of ChatGPT in Chinese and English settings, we designed an experiment evaluating its performance in the 2022 National Medical Licensing Examination (NMLE) in China. For a standardized evaluation, we used the comprehensive written part of the NMLE, translated into English by a bilingual expert. All questions were input into ChatGPT, which provided answers and reasons for choosing them. Responses were evaluated for "information quality" using the Likert scale.ResultsChatGPT demonstrated a correct response rate of 81.25% for Chinese and 86.25% for English questions. Logistic regression analysis showed that neither the difficulty nor the subject matter of the questions was a significant factor in AI errors. The Brier Scores, indicating predictive accuracy, were 0.19 for Chinese and 0.14 for English, indicating good predictive performance. The average quality score for English responses was excellent (4.43 point), slightly higher than for Chinese (4.34 point).ConclusionWhile AI language models like ChatGPT show promise for global healthcare, language bias is a key challenge. Ensuring that such technologies are robustly trained and sensitive to multiple languages and cultures is vital. Further research into AI's role in healthcare, particularly in areas with limited resources, is warranted.
引用
收藏
页数:7
相关论文
共 22 条
  • [1] ChatGPT Performs on the Chinese National Medical Licensing Examination
    Xinyi Wang
    Zhenye Gong
    Guoxin Wang
    Jingdan Jia
    Ying Xu
    Jialu Zhao
    Qingye Fan
    Shaun Wu
    Weiguo Hu
    Xiaoyang Li
    Journal of Medical Systems, 47
  • [2] ChatGPT Performs on the Chinese National Medical Licensing Examination
    Wang, Xinyi
    Gong, Zhenye
    Wang, Guoxin
    Jia, Jingdan
    Xu, Ying
    Zhao, Jialu
    Fan, Qingye
    Wu, Shaun
    Hu, Weiguo
    Li, Xiaoyang
    JOURNAL OF MEDICAL SYSTEMS, 2023, 47 (01)
  • [3] Accuracy of ChatGPT on Medical Questions in the National Medical Licensing Examination in Japan: Evaluation Study
    Yanagita, Yasutaka
    Yokokawa, Daiki
    Uchida, Shun
    Tawara, Junsuke
    Ikusaka, Masatomi
    JMIR FORMATIVE RESEARCH, 2023, 7
  • [4] Performance of ChatGPT on Chinese national medical licensing examinations: a five-year examination evaluation study for physicians, pharmacists and nurses
    Hui Zong
    Jiakun Li
    Erman Wu
    Rongrong Wu
    Junyu Lu
    Bairong Shen
    BMC Medical Education, 24
  • [5] Performance of ChatGPT on Chinese national medical licensing examinations: a five-year examination evaluation study for physicians, pharmacists and nurses
    Zong, Hui
    Li, Jiakun
    Wu, Erman
    Wu, Rongrong
    Lu, Junyu
    Shen, Bairong
    BMC MEDICAL EDUCATION, 2024, 24 (01)
  • [6] ChatGPT-4: An assessment of an upgraded artificial intelligence chatbot in the United States Medical Licensing Examination
    Mihalache, Andrew
    Huang, Ryan S.
    Popovic, Marko M.
    Muni, Rajeev H.
    MEDICAL TEACHER, 2024, 46 (03) : 366 - 372
  • [7] Can ChatGPT pass China's national medical licensing examination?
    Shang, Luxiang
    Xue, Mingyue
    Hou, Yinglong
    Tang, Baopeng
    ASIAN JOURNAL OF SURGERY, 2023, 46 (12) : 6112 - 6113
  • [8] Performance of DeepSeek-R1 and ChatGPT-4o on the Chinese National Medical Licensing Examination: A Comparative Study
    Jin Wu
    Zhiheng Wang
    Yifan Qin
    Journal of Medical Systems, 49 (1)
  • [9] Performance of ChatGPT-3.5 and ChatGPT-4 in the Taiwan National Pharmacist Licensing Examination: Comparative Evaluation Study
    Wang, Ying-Mei
    Shen, Hung-Wei
    Chen, Tzeng-Ji
    Chiang, Shu-Chiung
    Lin, Ting-Guan
    JMIR MEDICAL EDUCATION, 2025, 11
  • [10] Original Paper Performance of ChatGPT on the Peruvian National Licensing Medical Examination: Cross-Sectional Study
    Flores-Cohaila, Javier A.
    Garcia-Vicente, Abigail
    Vizcarra-Jimenez, Sonia F.
    De la Cruz-Galan, Janith
    Gutierrez-Arratia, Jesus
    Torres, Blanca Geraldine Quiroga
    Taype-Rondan, Alvaro
    JMIR MEDICAL EDUCATION, 2023, 9