Performance of ChatGPT 3.5 and 4 on U.S. dental examinations: the INBDE, ADAT, and DAT

被引:5
|
作者
Dashti, Mahmood [1 ]
Ghasemi, Shohreh [2 ]
Ghadimi, Niloofar [3 ]
Hefzi, Delband [4 ]
Karimian, Azizeh [5 ]
Zare, Niusha [6 ]
Fahimipour, Amir [7 ]
Khurshid, Zohaib [8 ]
Chafjiri, Maryam Mohammadalizadeh [9 ]
Ghaedsharaf, Sahar [10 ]
机构
[1] Shahid Beheshti Univ Med Sci, Res Inst Dent Sci, Dentofacial Deform Res Ctr, Tehran, Iran
[2] Queen Mary Coll, Dept Trauma & Craniofacial Reconstruct, London, England
[3] Islamic Azad Univ Med Sci, Dent Sch, Dept Oral & Maxillofacial Radiol, Tehran, Iran
[4] Univ Tehran Med Sci, Sch Dent, Tehran, Iran
[5] Golestan Univ Med Sci, Dent Res Ctr, Dept Biostat, Gorgan, Iran
[6] Univ Southern Calif, Dept Operat Dent, Los Angeles, CA USA
[7] Univ Sydney, Discipline Oral Surg Med & Diagnost, Sch Dent, Fac Med & Hlth,Westmead Ctr Oral Hlth, Sydney, Australia
[8] King Faisal Univ, Dept Prosthodont & Dent Implantol, Al Hasa, Saudi Arabia
[9] Shahid Beheshti Univ Med Sci, Sch Dent, Dept Oral & Maxillofacial Pathol, Tehran, Iran
[10] Shahid Beheshti Univ Med Sci, Sch Dent, Dept Oral & Maxillofacial Radiol, Tehran, Iran
关键词
Artificial Intelligence; Deep Learning; Dentistry; Education; Dental;
D O I
10.5624/isd.20240037
中图分类号
R78 [口腔科学];
学科分类号
1003 ;
摘要
Purpose: Recent advancements in artificial intelligence (AI), particularly tools such as ChatGPT developed by OpenAI, a U.S.-based AI research organization, have transformed the healthcare and education sectors. This study investigated the effectiveness of ChatGPT in answering dentistry exam questions, demonstrating its potential to enhance professional practice and patient care. Materials and Methods: This study assessed the performance of ChatGPT 3.5 and 4 on U.S. dental exams specifically, the Integrated National Board Dental Examination (INBDE), Dental Admission Test (DAT), and ChatGPT's answers were evaluated against official answer sheets. Results: ChatGPT 3.5 and 4 were tested with 253 questions from the INBDE, ADAT, and DAT exams. For the INBDE, both versions achieved 80% accuracy in knowledge-based questions and 66-69% in case history questions. In ADAT, they scored 66-83% in knowledge-based and 76% in case history questions. ChatGPT 4 excelled on the DAT, with 94% accuracy in knowledge-based questions, 57% in mathematical analysis items, and 100% in comprehension questions, surpassing ChatGPT 3.5's rates of 83%, 31%, and 82%, respectively. The difference was significant for knowledge-based questions (P= 0.009). Both versions showed similar patterns in incorrect responses. Conclusion: Both ChatGPT 3.5 and 4 effectively handled knowledge-based, case history, and comprehension questions, with ChatGPT 4 being more reliable and surpassing the performance of 3.5. ChatGPT 4's perfect score in comprehension questions underscores its trainability in specific subjects. However, both versions exhibited weaker performance in mathematical analysis, suggesting this as an area for improvement.
引用
收藏
页码:271 / 275
页数:5
相关论文
共 36 条
  • [1] Performance of ChatGPT-3.5 and ChatGPT-4o in the Japanese National Dental Examination
    Uehara, Osamu
    Morikawa, Tetsuro
    Harada, Fumiya
    Sugiyama, Nodoka
    Matsuki, Yuko
    Hiraki, Daichi
    Sakurai, Hinako
    Kado, Takashi
    Yoshida, Koki
    Murata, Yukie
    Matsuoka, Hirofumi
    Nagasawa, Toshiyuki
    Furuichi, Yasushi
    Abiko, Yoshihiro
    Miura, Hiroko
    JOURNAL OF DENTAL EDUCATION, 2024,
  • [2] Assessing the Performance of ChatGPT 3.5 and ChatGPT 4 in Operative Dentistry and Endodontics: An Exploratory Study
    Snigdha, Niher Tabassum
    Batul, Rumesa
    Karobari, Mohmed Isaqali
    Adil, Abdul Habeeb
    Dawasaz, Ali Azhar
    Hameed, Mohammad Shahul
    Mehta, Vini
    Noorani, Tahir Yusuf
    HUMAN BEHAVIOR AND EMERGING TECHNOLOGIES, 2024, 2024
  • [3] Performance of ChatGPT on Solving Orthopedic Board-Style Questions: A Comparative Analysis of ChatGPT 3.5 and ChatGPT 4
    Kim, Sung Eun
    Lee, Ji Han
    Choi, Byung Sun
    Han, Hyuk-Soo
    Lee, Myung Chul
    Ro, Du Hyun
    CLINICS IN ORTHOPEDIC SURGERY, 2024, 16 (04) : 669 - 673
  • [4] Performance of ChatGPT and GPT-4 on Neurosurgery Written Board Examinations
    Ali, Rohaid
    Tang, Oliver Y.
    Connolly, Ian D.
    Sullivan, Patricia L. Zadnik
    Shin, John H.
    Fridley, Jared S.
    Asaad, Wael F.
    Cielo, Deus
    Oyelese, Adetokunbo A.
    Doberstein, Curtis E.
    Gokaslan, Ziya L.
    Telfeian, Albert E.
    NEUROSURGERY, 2023, 93 (06) : 1353 - 1365
  • [5] Matching Human Expertise: ChatGPT's Performance on Hand Surgery Examinations
    Kirschenbaum, Zachary A.
    Han, Yuri
    Vrindten, Kiera L.
    Wang, Hanbin
    Cody, Ron
    Katt, Brian M.
    Kirschenbaum, David
    HAND-AMERICAN ASSOCIATION FOR HAND SURGERY, 2025,
  • [6] Pharmacy student use of ChatGPT: A survey of students at a U.S. School of Pharmacy
    Anderson, Heather D.
    Kwon, Sue
    Linnebur, Lauren A.
    Valdez, Connie A.
    Linnebur, Sunny A.
    CURRENTS IN PHARMACY TEACHING AND LEARNING, 2024, 16 (11)
  • [7] Performance of ChatGPT-3.5 and ChatGPT-4 on the European Board of Urology (EBU) exams: a comparative analysis
    Schoch, Justine
    Schmelz, H. -u.
    Strauch, Angelina
    Borgmann, Hendrik
    Nestler, Tim
    WORLD JOURNAL OF UROLOGY, 2024, 42 (01)
  • [8] Saudi and U.S. Dental Student Attitudes Toward Treating Individuals with Developmental Disabilities
    Alkahtani, Zuhair M.
    Stark, Paul C.
    Loo, Cheen Y.
    Wright, Wanda G.
    Morgan, John P.
    JOURNAL OF DENTAL EDUCATION, 2014, 78 (08) : 1145 - 1153
  • [9] Can ChatGPT-3.5 Pass a Medical Exam? A Systematic Review of ChatGPT's Performance in Academic Testing
    Sumbal, Anusha
    Sumbal, Ramish
    Amir, Alina
    JOURNAL OF MEDICAL EDUCATION AND CURRICULAR DEVELOPMENT, 2024, 11
  • [10] Artificial intelligence in dental education: ChatGPT's performance on the periodontic in-service examination
    Danesh, Arman
    Pazouki, Hirad
    Danesh, Farzad
    Danesh, Arsalan
    Vardar-Sengul, Saynur
    JOURNAL OF PERIODONTOLOGY, 2024, 95 (07) : 682 - 687