Evaluation of ChatGPT's Performance in the Turkish Board of Orthopaedic Surgery Examination

被引:0
|
作者
Yigitbay, Ahmet [1 ]
机构
[1] Siverek State Hosp, Clin Orthoped & Traumatol, Sanliurfa, Turkiye
来源
HASEKI TIP BULTENI-MEDICAL BULLETIN OF HASEKI | 2024年 / 62卷 / 04期
关键词
Artificial intelligence; humans; orthopedics; specialty boards; ARTIFICIAL-INTELLIGENCE;
D O I
10.4274/haseki.galenos.2024.10038
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Aim: Technological advances lead to significant changes in education and evaluation processes in medicine. In particular, artificial intelligence and natural language processing developments offer new opportunities in the health sector. This article evaluates Chat Generative Pre-Trained Transformer's (ChatGPT) performance in the Turkish Orthopaedics and Traumatology Education Council (TOTEK) Qualifying Written Examination and its applicability. Methods: To evaluate ChatGPT's performance, TOTEK Qualifying Written Examination questions from the last five years were entered as data. The results of ChatGPT were assessed under four parameters and compared with the actual exam results. The results were analyzed statistically. Results: Of the 500 questions, 458 were used as data in this study. Chat Generative Pre-Trained Transformer scored 40.2%, 26.3%, 37.3%, 32.9%, and 35.8% in the 2019, 2020, 2021, 2022, and 2023 TOTEK Qualifying Written Examination, respectively. When the correct answer percentages of ChatGPT according to years and the simple linear regression model applied to these data were analyzed, it was determined that there was a slightly decreasing trend in the correct answer rates as the years progressed. ChatGPT's TOTEK Qualifying Written Examination performance showed a statistically significant difference from the actual exam results. It was observed that the correct answer percentage of ChatGPT was below the general average success scores of the exam for each year. Conclusions: This analysis of artificial intelligence's applicability in the field and its role in training processes is essential to assess ChatGPT's potential uses and limitations. Chat Generative Pre-Trained Transformer can be a training tool, especially for knowledgebased and logical questions on specific topics. Still, its current performance is not at a level that can replace human decision-making in specialized medical fields.
引用
收藏
页码:243 / 249
页数:7
相关论文
共 50 条
  • [31] Reshaping medical education: Performance of ChatGPT on a PES medical examination
    Wojcik, Simona
    Rulkiewicz, Anna
    Pruszczyk, Piotr
    Lisik, Wojciech
    Pobozy, Marcin
    Domienik-Karlowicz, Justyna
    CARDIOLOGY JOURNAL, 2024, 31 (03) : 442 - 450
  • [32] Performance of ChatGPT on the National Korean Occupational Therapy Licensing Examination
    Lee, Si-An
    Heo, Seoyoon
    Park, Jin-Hyuck
    DIGITAL HEALTH, 2024, 10
  • [33] Evaluation of ChatGPT pathology knowledge using board-style questions
    Geetha, Saroja D.
    Khan, Anam
    Khan, Atif
    Kannadath, Bijun S.
    Vitkovski, Taisia
    AMERICAN JOURNAL OF CLINICAL PATHOLOGY, 2024, 161 (04) : 393 - 398
  • [34] Reshaping medical education: Performance of ChatGPT on a PES medical examination
    Wojcik, Simona
    Rulkiewicz, Anna
    Pruszczyk, Piotr
    Lisik, Wojciech
    Pobozy, Marcin
    Domienik-Karlowicz, Justyna
    CARDIOLOGY JOURNAL, 2023, : 442 - 450
  • [35] Performance of ChatGPT on the Brazilian Radiology and Diagnostic Imaging and Mammography Board Examinations
    Almeida, Leonardo C.
    Farina, Eduardo M. J. M.
    Kurilei, Paulo E. A.
    Abdala, Nitamar
    Kitamura, Felipe C.
    RADIOLOGY-ARTIFICIAL INTELLIGENCE, 2024, 6 (01)
  • [36] Performance of ChatGPT-3.5 and ChatGPT-4 on the European Board of Urology (EBU) exams: a comparative analysis
    Schoch, Justine
    Schmelz, H. -u.
    Strauch, Angelina
    Borgmann, Hendrik
    Nestler, Tim
    WORLD JOURNAL OF UROLOGY, 2024, 42 (01)
  • [37] Assessment of ChatGPT success with specialty medical knowledge using anaesthesiology board examination practice questions
    Shay, Denys
    Kumar, Bhawesh
    Bellamy, David
    Palepu, Anil
    Dershwitz, Mark
    Walz, Jens M.
    Schaefer, Maximilian S.
    Beam, Andrew
    BRITISH JOURNAL OF ANAESTHESIA, 2023, 131 (02)
  • [38] Performance of ChatGPT and Dental Students on Concepts of Periodontal Surgery
    Li, Chen
    Zhang, Jinmei
    Abdul-Masih, John
    Zhang, Sihan
    Yang, Jingmei
    EUROPEAN JOURNAL OF DENTAL EDUCATION, 2025, 29 (01) : 36 - 43
  • [39] Performance evaluation of ChatGPT-4.0 and Gemini on image-based neurosurgery board practice questions: A comparative analysis
    Mcnulty, Alana M.
    Valluri, Harshitha
    Gajjar, Avi A.
    Custozzo, Amanda
    Field, Nicholas C.
    Paul, Alexandra R.
    JOURNAL OF CLINICAL NEUROSCIENCE, 2025, 134
  • [40] Impact of Attached File Formats on the Performance of ChatGPT-4 on the Japanese National Nursing Examination: Evaluation Study
    Taira, Kazuya
    Itaya, Takahiro
    Yada, Shuntaro
    Hiyama, Kirara
    Hanada, Ayame
    JMIR NURSING, 2025, 8