ChatGPT Knowledge Evaluation in Basic and Clinical Medical Sciences: Multiple Choice Question Examination-Based Performance

被引:40
作者
Meo, Sultan Ayoub [1 ]
Al-Masri, Abeer A. [1 ]
Alotaibi, Metib [2 ]
Meo, Muhammad Zain Sultan [3 ]
Meo, Muhammad Omair Sultan [3 ]
机构
[1] King Saud Univ, Coll Med, Dept Physiol, Riyadh 11461, Saudi Arabia
[2] King Saud Univ, Univ Diabet Unit, Coll Med, Dept Med, Riyadh 11461, Saudi Arabia
[3] Alfaisal Univ, Coll Med, Riyadh 11533, Saudi Arabia
关键词
ChatGPT; knowledge; intellect level; medical education;
D O I
10.3390/healthcare11142046
中图分类号
R19 [保健组织与事业(卫生事业管理)];
学科分类号
摘要
The Chatbot Generative Pre-Trained Transformer (ChatGPT) has garnered great attention from the public, academicians and science communities. It responds with appropriate and articulate answers and explanations across various disciplines. For the use of ChatGPT in education, research and healthcare, different perspectives exist with some level of ambiguity around its acceptability and ideal uses. However, the literature is acutely lacking in establishing a link to assess the intellectual levels of ChatGPT in the medical sciences. Therefore, the present study aimed to investigate the knowledge level of ChatGPT in medical education both in basic and clinical medical sciences, multiple-choice question (MCQs) examination-based performance and its impact on the medical examination system. In this study, initially, a subject-wise question bank was established with a pool of multiple-choice questions (MCQs) from various medical textbooks and university examination pools. The research team members carefully reviewed the MCQ contents and ensured that the MCQs were relevant to the subject's contents. Each question was scenario-based with four sub-stems and had a single correct answer. In this study, 100 MCQs in various disciplines, including basic medical sciences (50 MCQs) and clinical medical sciences (50 MCQs), were randomly selected from the MCQ bank. The MCQs were manually entered one by one, and a fresh ChatGPT session was started for each entry to avoid memory retention bias. The task was given to ChatGPT to assess the response and knowledge level of ChatGPT. The first response obtained was taken as the final response. Based on a pre-determined answer key, scoring was made on a scale of 0 to 1, with zero representing incorrect and one representing the correct answer. The results revealed that out of 100 MCQs in various disciplines of basic and clinical medical sciences, ChatGPT attempted all the MCQs and obtained 37/50 (74%) marks in basic medical sciences and 35/50 (70%) marks in clinical medical sciences, with an overall score of 72/100 (72%) in both basic and clinical medical sciences. It is concluded that ChatGPT obtained a satisfactory score in both basic and clinical medical sciences subjects and demonstrated a degree of understanding and explanation. This study's findings suggest that ChatGPT may be able to assist medical students and faculty in medical education settings since it has potential as an innovation in the framework of medical sciences and education.
引用
收藏
页数:11
相关论文
共 32 条
[1]  
Al-Rukban MO, 2006, J FAM COMMUNITY MED, V13, P125
[2]   Evaluating the effectiveness of 'MCQ development workshop using cognitive model framework: A pre-post study [J].
Ali, Rahila ;
Sultan, Amber Shamim ;
Zahid, Nida .
JOURNAL OF THE PAKISTAN MEDICAL ASSOCIATION, 2021, 71 (01) :119-121
[3]   Assessing the Capability of ChatGPT in Answering First- and Second-Order Knowledge Questions on Microbiology as per Competency- Based Medical Education Curriculum [J].
Das, Dipmala ;
Kumar, Nikhil ;
Longjam, Langamba Angom ;
Sinha, Ranwir ;
Roy, Asitava Deb ;
Mondal, Himel ;
Gupta, Pratima .
CUREUS JOURNAL OF MEDICAL SCIENCE, 2023, 15 (03)
[4]  
Duong D, 2024, EUR J HUM GENET, V32, P466, DOI 10.1038/s41431-023-01396-8
[5]   Evaluating ChatGPT's Ability to Solve Higher-Order Questions on the Competency-Based Medical Education Curriculum in Medical Biochemistry [J].
Ghosh, Arindam ;
Bir, Aritri .
CUREUS JOURNAL OF MEDICAL SCIENCE, 2023, 15 (04)
[6]   How Does ChatGPT Perform on the United States Medical Licensing Examination (USMLE)? The Implications of Large Language Models for Medical Education and Knowledge Assessment [J].
Gilson, Aidan ;
Safranek, Conrad W. ;
Huang, Thomas ;
Socrates, Vimig ;
Chi, Ling ;
Taylor, Richard Andrew ;
Chartash, David .
JMIR MEDICAL EDUCATION, 2023, 9
[7]   Medical students create multiple-choice questions for learning in pathology education: a pilot study [J].
Grainger, Rebecca ;
Dai, Wei ;
Osborne, Emma ;
Kenwright, Diane .
BMC MEDICAL EDUCATION, 2018, 18
[8]   Performance of ChatGPT on the Plastic Surgery Inservice Training Examination [J].
Gupta, Rohun ;
Herzog, Isabel ;
Park, John B. ;
Weisberger, Joseph ;
Firouzbakht, Peter ;
Ocon, Vanessa ;
Chao, John ;
Lee, Edward S. ;
Mailey, Brian A. .
AESTHETIC SURGERY JOURNAL, 2023, :NP1078-NP1082
[9]  
Ha LA, 2019, ACL Anthology, P418
[10]   An exploratory survey about using ChatGPT in education, healthcare, and research [J].
Hosseini, Mohammad ;
Gao, Catherine A. ;
Liebovitz, David M. ;
Carvalho, Alexandre M. ;
Ahmad, Faraz S. ;
Luo, Yuan ;
MacDonald, Ngan ;
Holmes, Kristi L. ;
Kho, Abel .
PLOS ONE, 2023, 18 (10)