ChatGPT to generate clinical vignettes for teaching and multiple-choice questions for assessment: A randomized controlled experiment

被引:21
作者
Coskun, Oezlem [1 ,2 ]
Kiyak, Yavuz Selim [1 ]
Budakoglu, Isil Irem [1 ]
机构
[1] Gazi Univ, Dept Med Educ & Informat, Ankara, Turkiye
[2] Gazi Univ, Dept Med Educ & Informat, Hastanesi E Blok 9 Kat Besevler, Ankara, Turkiye
关键词
ChatGPT; artificial intelligence; automatic item generation; clinical vignette; medical education;
D O I
10.1080/0142159X.2024.2327477
中图分类号
G40 [教育学];
学科分类号
040101 ; 120403 ;
摘要
Aim: This study aimed to evaluate the real-life performance of clinical vignettes and multiple-choice questions generated by using ChatGPT. Methods: This was a randomized controlled study in an evidence-based medicine training program. We randomly assigned seventy-four medical students to two groups. The ChatGPT group received ill-defined cases generated by ChatGPT, while the control group received human-written cases. At the end of the training, they evaluated the cases by rating 10 statements using a Likert scale. They also answered 15 multiple-choice questions (MCQs) generated by ChatGPT. The case evaluations of the two groups were compared. Some psychometric characteristics (item difficulty and point-biserial correlations) of the test were also reported. Results: None of the scores in 10 statements regarding the cases showed a significant difference between the ChatGPT group and the control group (p > .05). In the test, only six MCQs had acceptable levels (higher than 0.30) of point-biserial correlation, and five items could be considered acceptable in classroom settings. Conclusions: The results showed that the quality of the vignettes are comparable to those created by human authors, and some multiple-questions have acceptable psychometric characteristics. ChatGPT has potential in generating clinical vignettes for teaching and MCQs for assessment in medical education.
引用
收藏
页码:268 / 274
页数:7
相关论文
共 41 条
[1]  
Aje AA, 2024, BMC ENDOCR DISORD, V24, DOI [10.1186/s12902-024-01696-0, 10.1080/14703297.2023.2190148]
[2]   Sailing the Seven Seas: A Multinational Comparison of ChatGPT's Performance on Medical Licensing Examinations [J].
Alfertshofer, Michael ;
Hoch, Cosima C. ;
Funk, Paul F. ;
Hollmann, Katharina ;
Wollenberg, Barbara ;
Knoedler, Samuel ;
Knoedler, Leonard .
ANNALS OF BIOMEDICAL ENGINEERING, 2024, 52 (06) :1542-1545
[3]   Using artificial intelligence to create diverse and inclusive medical case vignettes for education [J].
Bakkum, Michiel J. ;
Hartjes, Marielle G. ;
Piet, Joost D. ;
Donker, Erik M. ;
Likic, Robert ;
Sanz, Emilio ;
de Ponti, Fabrizio ;
Verdonk, Petra ;
Richir, Milan C. ;
van Agtmael, Michiel A. ;
Tichelaar, Jelle .
BRITISH JOURNAL OF CLINICAL PHARMACOLOGY, 2024, 90 (03) :640-648
[4]  
Benoit JR., 2023, MEDRXIV
[5]  
Brown TB, 2020, ADV NEUR IN, V33
[6]  
Buchholz K., 2023, STATISTA INFOGRAPHIC
[7]   Is "ChatGPT" capable of passing the 2022 MIR exam? Implications of artificial intelligence in medical education in Spain [J].
Carrasco, J. P. ;
Garcia, E. ;
Sanchez, D. A. ;
Porter, P. D. Estrella ;
De La Puente, L. ;
Navarro, J. ;
Cerame, A. .
SPANISH JOURNAL OF MEDICAL EDUCATION, 2023, 4 (01)
[8]   ChatGPT versus human in generating medical graduate exam multiple choice questions-A multinational prospective study (Hong Kong SAR, Singapore, Ireland, and the United Kingdom) [J].
Cheung, Billy Ho Hung ;
Lau, Gary Kui Kai ;
Wong, Gordon Tin Chun ;
Lee, Elaine Yuen Phin ;
Kulkarni, Dhananjay ;
Seow, Choon Sheong ;
Wong, Ruby ;
Co, Michael Tiong-Hong .
PLOS ONE, 2023, 18 (08)
[9]   Chatting or cheating? The impacts of ChatGPT and other artificial intelligence language models on nurse education [J].
Choi, Edmond Pui Hang ;
Lee, Jung Jae ;
Ho, Mu-Hsing ;
Kwok, Jojo Yan Yan ;
Lok, Kris Yuet Wan .
NURSE EDUCATION TODAY, 2023, 125
[10]   A Novel Approach to Teach Evidence-Based Medicine: Modified PEARLS [J].
Coskun, Ozlem ;
Kiyak, Yavuz Selim ;
Budakoglu, Isil Irem ;
Ergun, Mehmet Ali ;
Haznedaroglu, Seminur .
GAZI MEDICAL JOURNAL, 2022, 33 (04) :329-336