Could ChatGPT Pass the UK Radiology Fellowship Examinations?

被引:19
作者
Ariyaratne, Sisith [1 ]
Jenko, Nathan [1 ]
Davies, A. Mark [1 ]
Iyengar, Karthikeyan P. [2 ]
Botchu, Rajesh [1 ]
机构
[1] Royal Orthopaed Hosp NHS Fdn Trust, Dept Musculoskeletal Radiol, Birmingham, England
[2] Mersey & West Lancashire Teaching NHS Trust, Southport & Ormskirk Hosp, Dept Trauma & Orthopaed, Southport, England
关键词
FRCR examination; ChatGPT; Artificial intelligence;
D O I
10.1016/j.acra.2023.11.026
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Rationale and Objectives: Chat Generative Pre-trained Transformer (ChatGPT) is an artificial intelligence (AI) tool which utilises machine learning to generate original text resembling human language. AI models have recently demonstrated remarkable ability at analysing and solving problems, including passing professional examinations. We investigate the performance of ChatGPT on some of the UK radiology fellowship equivalent examination questions. Methods: ChatGPT was asked to answer questions from question banks resembling the Fellowship of the Royal College of Radiologists (FRCR) examination. The entire physics part 1 question bank (203 5 -part true/false questions) was answered by the GPT-4 model and answers recorded. 240 single best answer questions (SBAs) (representing the true length of the FRCR 2A examination) were answered by both GPT-3.5 and GPT-4 models. Results: ChatGPT 4 answered 74.8% of part 1 true/false statements correctly. The spring 2023 passing mark of the part 1 examination was 75.5% and ChatGPT thus narrowly failed. In the 2A examination, ChatGPT 3.5 answered 50.8% SBAs correctly, while GPT-4 answered 74.2% correctly. The winter 2022 2A pass mark was 63.3% and thus GPT-4 clearly passed. Conclusion: AI models such as ChatGPT are able to answer the majority of questions in an FRCR style examination. It is reasonable to assume that further developments in AI will be more likely to succeed in comprehending and solving questions related to medicine, specifically clinical radiology. Advances in knowledge: Our findings outline the unprecedented capabilities of AI, adding to the current relatively small body of literature on the subject, which in turn can play a role medical training, evaluation and practice. This can undoubtedly have implications for radiology.
引用
收藏
页码:2178 / 2182
页数:5
相关论文
共 11 条
[1]   Will collaborative publishing with ChatGPT drive academic writing in the future? [J].
Ariyaratne, Sisith ;
Iyengar, Kathikeyan P. ;
Botchu, Rajesh .
BRITISH JOURNAL OF SURGERY, 2023, 110 (09) :1213-1214
[2]   ChatGPT in academic publishing: An ally or an adversary? [J].
Ariyaratne, Sisith ;
Botchu, Rajesh ;
Iyengar, Karthikeyan P. .
SCOTTISH MEDICAL JOURNAL, 2023, 68 (03) :129-130
[3]   A comparison of ChatGPT-generated articles with human-written articles [J].
Ariyaratne, Sisith ;
Iyengar, Karthikeyan. P. ;
Nischal, Neha ;
Chitti Babu, Naparla ;
Botchu, Rajesh .
SKELETAL RADIOLOGY, 2023, 52 (09) :1755-1758
[4]   Potential Applications and Impact of ChatGPT in Radiology [J].
Bajaj, Suryansh ;
Gandhi, Darshan ;
Nayar, Divya .
ACADEMIC RADIOLOGY, 2024, 31 (04) :1256-1261
[5]   Performance of ChatGPT on a Radiology Board-style Examination: Insights into Current Strengths and Limitations [J].
Bhayana, Rajesh ;
Krishna, Satheesh ;
Bleakney, Robert R. .
RADIOLOGY, 2023, 307 (05)
[6]  
Hassankhani A, 2023, Acad Radiol, P00279
[7]   Assessing AI-Powered Patient Education: A Case Study in Radiology [J].
Kuckelman, Ian J. ;
Yi, Paul H. ;
Bui, Molinna ;
Onuh, Ifeanyi ;
Anderson, Jade A. ;
Ross, Andrew B. .
ACADEMIC RADIOLOGY, 2024, 31 (01) :338-342
[8]  
Kung TH, 2023, PLoS digital health, V2, DOI DOI 10.1371/JOURNALPDIG.0000198
[9]  
Nori H, 2023, Arxiv, DOI [arXiv:2303.13375, DOI 10.48550/ARXIV.2303.13375, 10.48550/arXiv.2303.13375]
[10]  
OpenAI, Introducing ChatGPT