The performance of artificial intelligence language models in board-style dental knowledge assessment A preliminary study on ChatGPT

被引:18
作者
Danesh, Arman [1 ,5 ]
Pazouki, Hirad [2 ]
Danesh, Kasra [3 ]
Danesh, Farzad
Danesh, Arsalan [4 ]
机构
[1] Western Univ, Schulich Sch Med & Dent, London, ON, Canada
[2] Western Univ, Fac Hlth Sci, London, ON, Canada
[3] Florida Atlantic Univ, Coll Engn & Comp Sci, Boca Raton, FL 33431 USA
[4] Nova Southeastern Univ, Coll Dent Med, Dept Periodontol, Ft Lauderdale, FL 33314 USA
[5] Nova Southeastern Univ, Coll Dent Med, Dept Oral & Maxillofacial Surg, 3200 S Univ Dr, Davie, FL 33328 USA
关键词
Artificial intelligence; ChatGPT; dental board examination; dental education; dentistry; Integrated National Board Dental Examination;
D O I
10.1016/j.adaj.2023.07.016
中图分类号
R78 [口腔科学];
学科分类号
1003 ;
摘要
Background. Although Chat Generative Pre-trained Transformer (ChatGPT) (OpenAI) may be an appealing educational resource for students, the chatbot responses can be subject to misinfor-mation. This study was designed to evaluate the performance of ChatGPT on a board-style mul-tiple-choice dental knowledge assessment to gauge its capacity to output accurate dental content and in turn the risk of misinformation associated with use of the chatbot as an educational resource by dental students.Methods. ChatGPT3.5 and ChatGPT4 were asked questions obtained from 3 different sources: INBDE Bootcamp, ITDOnline, and a list of board-style questions provided by the Joint Commis-sion on National Dental Examinations. Image-based questions were excluded, as ChatGPT only takes text-based inputs. The mean performance across 3 trials was reported for each model.Results. ChatGPT3.5 and ChatGPT4 answered 61.3% and 76.9% of the questions correctly on average, respectively. A 2-tailed t test was used to compare 2 independent sample means, and a 2-tailed c2 test was used to compare 2 sample proportions. A P value less than .05 was considered to be statistically significant.Conclusion. ChatGPT3.5 did not perform sufficiently well on the board-style knowledge assess-ment. ChatGPT4, however, displayed a competent ability to output accurate dental content. Future research should evaluate the proficiency of emerging models of ChatGPT in dentistry to assess its evolving role in dental education. Practical Implications. Although ChatGPT showed an impressive ability to output accurate dental content, our findings should encourage dental students to incorporate ChatGPT to sup-plement their existing learning program instead of using it as their primary learning resource.
引用
收藏
页码:970 / 974
页数:5
相关论文
共 50 条
  • [21] Performance of artificial intelligence chatbots in sleep medicine certification board exams: ChatGPT versus Google Bard
    Cheong, Ryan Chin Taw
    Pang, Kenny Peter
    Unadkat, Samit
    Mcneillis, Venkata
    Williamson, Andrew
    Joseph, Jonathan
    Randhawa, Premjit
    Andrews, Peter
    Paleri, Vinidh
    EUROPEAN ARCHIVES OF OTO-RHINO-LARYNGOLOGY, 2024, 281 (04) : 2137 - 2143
  • [22] Generative Artificial Intelligence Through ChatGPT and Other Large Language Models in Ophthalmology Clinical Applications and Challenges
    Tan, Ting Fang
    Thirunavukarasu, Arun James
    Campbell, J. Peter
    Keane, Pearse A.
    Pasquale, Louis R.
    Abramoff, Michael D.
    Kalpathy-Cramer, Jayashree
    Lum, Flora
    Kim, Judy E.
    Baxter, Sally L.
    Ting, Daniel Shu Wei
    OPHTHALMOLOGY SCIENCE, 2023, 3 (04):
  • [23] Artificial intelligence chatbots and large language models in dental education: Worldwide survey of educators
    Uribe, Sergio E.
    Maldupa, Ilze
    Kavadella, Argyro
    El Tantawi, Maha
    Chaurasia, Akhilanand
    Fontana, Margherita
    Marino, Rodrigo
    Innes, Nicola
    Schwendicke, Falk
    EUROPEAN JOURNAL OF DENTAL EDUCATION, 2024, 28 (04) : 865 - 876
  • [24] Artificial intelligence and social intelligence: preliminary comparison study between AI models and psychologists
    Sufyan, Nabil Saleh
    Fadhel, Fahmi H.
    Alkhathami, Saleh Safeer
    Mukhadi, Jubran Y. A.
    FRONTIERS IN PSYCHOLOGY, 2024, 15
  • [25] Assessment of knowledge and awareness of artificial intelligence and its uses in dentistry among dental students
    Vamshi Ram, V.
    Sadeep, Hima
    JOURNAL OF PHARMACEUTICAL NEGATIVE RESULTS, 2022, 13 : 1304 - 1309
  • [26] Performance of artificial intelligence on Turkish dental specialization exam: can ChatGPT-4.0 and gemini advanced achieve comparable results to humans?
    Sismanoglu, Soner
    Capan, Belen Sirinoglu
    BMC MEDICAL EDUCATION, 2025, 25 (01)
  • [27] Making sense of artificial intelligence and large language models-including ChatGPT-in pediatric hematology/oncology
    Wyatt, Kirk D.
    Alexander, Natasha
    Hills, Gerard D.
    Liang, Wayne H.
    Kadauke, Stephan
    Volchenboum, Samuel L.
    Mian, Amir
    Phillips, Charles A.
    PEDIATRIC BLOOD & CANCER, 2024, 71 (09)
  • [28] Best practices for implementing ChatGPT, large language models, and artificial intelligence in qualitative and survey-based research
    Kantor, Jonathan
    JAAD INTERNATIONAL, 2024, 14 : 22 - 23
  • [29] Assessment of Artificial Intelligence Platforms With Regard to Medical Microbiology Knowledge: An Analysis of ChatGPT and Gemini
    Ranjan, Jai
    Ahmad, Absar
    Subudhi, Monalisa
    Kumar, Ajay
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2024, 16 (05)
  • [30] Performance of the Large Language Models in African rheumatology: a diagnostic test accuracy study of ChatGPT-4, Gemini, Copilot, and Claude artificial intelligence
    Yannick Laurent Tchenadoyo Bayala
    Wendlassida Joelle Stéphanie Zabsonré/Tiendrebeogo
    Dieu-Donné Ouedraogo
    Fulgence Kaboré
    Charles Sougué
    Aristide Relwendé Yameogo
    Wendlassida Martin Nacanabo
    Ismael Ayouba Tinni
    Aboubakar Ouedraogo
    Yamyellé Enselme Zongo
    BMC Rheumatology, 9 (1)