The Emotional Intelligence of the GPT-4 Large Language Model

被引:2
|
作者
Vzorin, Gleb D. [1 ,2 ]
Bukinich, Alexey M. [1 ,3 ]
Sedykh, Anna V. [1 ]
Vetrova, Irina I. [2 ]
Sergienko, Elena A. [2 ]
机构
[1] Lomonosov Moscow State Univ, Moscow, Russia
[2] Russian Acad Sci, Inst Psychol, Moscow, Russia
[3] Fed Sci Ctr Psychol & Interdisciplinary Res, Moscow, Russia
来源
PSYCHOLOGY IN RUSSIA-STATE OF THE ART | 2024年 / 17卷 / 02期
关键词
artificial empathy; artificial psychology; ChatGPT; emotional intelligence (EI); emotional quotient(EQ); GPT-4; machine behavior;
D O I
10.11621/pir.2024.0206
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Background. Advanced AI models such as the large language model GPT-4 demonstrate sophisticated intellectual capabilities, sometimes exceeding human intellectual performance. However, the emotional competency of these models, along with their underlying mechanisms, has not been sufficiently evaluated. Objective. Our research aimed to explore different emotional intelligence domains in GPT-4 according to the Mayer-Salovey-Caruso model. We also tried to find out whether GPT-4's answer accuracy is consistent with its explanation of the answer. Design. The Russian version of the Mayer-Salovey-Caruso Emotional Intelligence Test (MSCEIT) sections was used in this research, with questions asked as text prompts in separate, independent ChatGPT chats three times each. Results. . High scores were achieved by the GPT-4 Large Language Model on the Understanding Emotions scale (with scores of 117, 124, and 128 across the three runs) and the Strategic Emotional Intelligence scale (with scores of 118, 121, and 122). Average scores were obtained on the Managing Emotions scale (103, 108, and 110 points). However, the Using Emotions to Facilitate Thought scale yielded low and less reliable scores (85, 86, and 88 points). Four types of explanations for the answer choices were identified: Meaningless sentences; Relation declaration; Implicit logic; and Explicit logic. Correct answers were accompanied by all types of explanations, whereas incorrect answers were only followed by Meaningless sentences or Explicit logic. This distribution aligns with observed patterns in children when they explore and elucidate mental states. Conclusion. GPT-4 is capable of emotion identification and managing emotions, but it lacks deep reflexive analysis of emotional experience and the motivational aspect of emotions.
引用
收藏
页码:85 / 99
页数:15
相关论文
共 50 条
  • [31] The model student: GPT-4 performance on graduate biomedical science exams
    Stribling, Daniel
    Xia, Yuxing
    Amer, Maha K.
    Graim, Kiley S.
    Mulligan, Connie J.
    Renne, Rolf
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [32] Large Language Models as AI-Powered Educational Assistants: Comparing GPT-4 and Gemini for Writing Teaching Cases
    Lang, Guido
    Triantoro, Tamilla
    Sharp, Jason H.
    Journal of Information Systems Education, 35 (03): : 390 - 407
  • [33] On the Use of Large Language Models at Solving Math Problems: A Comparison Between GPT-4, LlaMA-2 and Gemini
    Navarro, Alejandro L. Garcia
    Koneva, Nataliia
    Hernandez, Jose Alberto
    Sanchez-Macian, Alfonso
    INTERNATIONAL JOURNAL OF INTERACTIVE MULTIMEDIA AND ARTIFICIAL INTELLIGENCE, 2025,
  • [34] Artificial Intelligence-Powered Hand Surgery Consultation: GPT-4 as an Assistant in a Hand Surgery Outpatient Clinic
    Leypold, Tim
    Schaefer, Benedikt
    Boos, Anja M.
    Beier, Justus P.
    JOURNAL OF HAND SURGERY-AMERICAN VOLUME, 2024, 49 (11): : 1078 - 1088
  • [35] On the Use of GPT-4 for Creating Goal Models: An Exploratory Study
    Chen, Boqi
    Chen, Kua
    Hassani, Shabnam
    Yang, Yujing
    Amyot, Daniel
    Lessard, Lysanne
    Mussbachcr, Gunter
    Sabetzadeh, Mehrdad
    Varro, Daniel
    2023 IEEE 31ST INTERNATIONAL REQUIREMENTS ENGINEERING CONFERENCE WORKSHOPS, REW, 2023, : 262 - 271
  • [36] Performance of GPT-4 Vision on kidney pathology exam questions
    Miao, Jing
    Thongprayoon, Charat
    Cheungpasitporn, Wisit
    Cornell, Lynn D.
    AMERICAN JOURNAL OF CLINICAL PATHOLOGY, 2024, 162 (03) : 220 - 226
  • [37] Performance of GPT-3.5 and GPT-4 on the Korean Pharmacist Licensing Examination: Comparison Study
    Jin, Hye Kyung
    Kim, Eunyoung
    JMIR MEDICAL EDUCATION, 2024, 10
  • [38] Performance of ChatGPT and GPT-4 on Neurosurgery Written Board Examinations
    Ali, Rohaid
    Tang, Oliver Y.
    Connolly, Ian D.
    Sullivan, Patricia L. Zadnik
    Shin, John H.
    Fridley, Jared S.
    Asaad, Wael F.
    Cielo, Deus
    Oyelese, Adetokunbo A.
    Doberstein, Curtis E.
    Gokaslan, Ziya L.
    Telfeian, Albert E.
    NEUROSURGERY, 2023, 93 (06) : 1353 - 1365
  • [39] Using artificial intelligence for exercise prescription in personalised health promotion: A critical evaluation of OpenAI's GPT-4 model
    Dergaa, Ismail
    Ben Saad, Helmi
    El Omri, Abdelfatteh
    Glenn, Jordan M.
    Clark, Cain C. T.
    Washif, Jad Adrian
    Guelmami, Noomen
    Hammouda, Omar
    Al-Horani, Ramzi A.
    Reynoso-Sanchez, Luis Felipe
    Romdhani, Mohamed
    Paineiras-Domingos, Laisa Liane
    Vancini, Rodrigo L.
    Taheri, Morteza
    Jose Mataruna-Dos-Santos, Leonardo
    Trabelsi, Khaled
    Chtourou, Hamdi
    Zghibi, Makram
    Eken, Ozgur
    Swed, Sarya
    Ben Aissa, Mohamed
    Shawki, Hossam H.
    El-Seedi, Hesham R.
    Mujika, Inigo
    Seiler, Stephen
    Zmijewski, Piotr
    Pyne, David B.
    Knechtle, Beat
    Asif, Irfan M.
    Drezner, Jonathan A.
    Sandbakk, Oyvind
    Chamari, Karim
    BIOLOGY OF SPORT, 2024, 41 (02) : 221 - 241
  • [40] Performance of GPT-4 on Chinese Nursing Examination
    Miao, Yiqun
    Luo, Yuan
    Zhao, Yuhan
    Li, Jiawei
    Liu, Mingxuan
    Wang, Huiying
    Chen, Yuling
    Wu, Ying
    NURSE EDUCATOR, 2024, 49 (06) : E338 - E343