Performance of ChatGPT on the Plastic Surgery Inservice Training Examination

被引:61
作者
Gupta, Rohun [1 ,3 ]
Herzog, Isabel [2 ]
Park, John B. [2 ]
Weisberger, Joseph
Firouzbakht, Peter [1 ]
Ocon, Vanessa
Chao, John [2 ]
Lee, Edward S.
Mailey, Brian A.
机构
[1] St Louis Univ, Dept Surg, Div Plast Surg, Sch Med, St Louis, MO USA
[2] Rutgers New Jersey Sch Med, Dept Plast Surg, Newark, NJ USA
[3] SLUCare Acad Pavil, 1008 S Spring Ave,Suite,1500 St Louis, St Louis, MO 63110 USA
关键词
D O I
10.1093/asj/sjad128
中图分类号
R61 [外科手术学];
学科分类号
摘要
Background Developed originally as a tool for resident self-evaluation, the Plastic Surgery Inservice Training Examination (PSITE) has become a standardized tool adopted by Plastic Surgery residency programs. The introduction of large language models (LLMs), such as ChatGPT (OpenAI, San Francisco, CA), has demonstrated the potential to help propel the field of Plastic Surgery. Objectives The authors of this study wanted to assess whether or not ChatGPT could be utilized as a tool in resident education by assessing its accuracy on the PSITE. Methods Questions were obtained from the 2022 PSITE, which was present on the American Council of Academic Plastic Surgeons (ACAPS) website. Questions containing images or tables were carefully inspected and flagged before being inputted into ChatGPT. All responses by ChatGPT were qualified utilizing the properties of natural coherence. Responses that were found to be incorrect were divided into the following categories: logical, informational, or explicit fallacy. Results ChatGPT answered a total of 242 questions with an accuracy of 54.96%. The software incorporated logical reasoning in 88.8% of questions, internal information in 95.5% of questions, and external information in 92.1% of questions. When stratified by correct and incorrect responses, we determined that there was a statistically significant difference in ChatGPT's use of external information (P < .05). Conclusions ChatGPT is a versatile tool that has the potential to impact resident education by providing general knowledge, clarifying information, providing case-based learning, and promoting evidence-based medicine. With advancements in LLM and artificial intelligence (AI), it is possible that ChatGPT may be an impactful tool for resident education within Plastic Surgery.
引用
收藏
页码:NP1078 / NP1082
页数:5
相关论文
共 16 条
[1]  
American Council of Academic Plastic Surgeons, US
[2]  
American Society of Plastic Surgeons, INS EX RES
[3]  
[Anonymous], ACGME Program Requirements for Graduate Medical Education in Internal Medicine
[4]  
Brown TB, 2020, ADV NEUR IN, V33
[5]  
Carlini N, 2021, PROCEEDINGS OF THE 30TH USENIX SECURITY SYMPOSIUM, P2633
[6]   Performance on the Plastic Surgery In-Service Examination Can Predict Success on the American Board of Plastic Surgery Written Examination [J].
Girotto, John A. ;
Adams, Nicholas S. ;
Janis, Jeffrey E. ;
Brandt, Keith E. ;
Slezak, Sheri S. .
PLASTIC AND RECONSTRUCTIVE SURGERY, 2019, 143 (05) :1099E-1105E
[7]   Response to: Truth or Lies? The Pitfalls and Limitations of ChatGPT in Systematic Review Creation [J].
Gupta, Rohun ;
Bagdady, Kazimir ;
Mailey, Brian A. .
AESTHETIC SURGERY JOURNAL, 2023, 43 (08) :NP656-NP657
[8]   Utilization of ChatGPT for Plastic Surgery Research: Friend or Foe? [J].
Gupta, Rohun ;
Herzog, Isabel ;
Weisberger, Joseph ;
Chao, John ;
Chaiyasate, Kongkrit ;
Lee, Edward S. .
JOURNAL OF PLASTIC RECONSTRUCTIVE AND AESTHETIC SURGERY, 2023, 80 :145-147
[9]   Expanding Cosmetic Plastic Surgery Research With ChatGPT [J].
Gupta, Rohun ;
Park, John B. ;
Bisht, Chirag ;
Herzog, Isabel ;
Weisberger, Joseph ;
Chao, John ;
Chaiyasate, Kongkrit ;
Lee, Edward S. .
AESTHETIC SURGERY JOURNAL, 2023, 43 (08) :930-937
[10]   Application of ChatGPT in Cosmetic Plastic Surgery: Ally or Antagonist? [J].
Gupta, Rohun ;
Pande, Parna ;
Herzog, Isabel ;
Weisberger, Joseph ;
Chao, John ;
Chaiyasate, Kongkrit ;
Lee, Edward S. .
AESTHETIC SURGERY JOURNAL, 2023, 43 (07) :NP587-NP590