The Emotional Intelligence of the GPT-4 Large Language Model

被引:1
|
作者
Vzorin, Gleb D. [1 ,2 ]
Bukinich, Alexey M. [1 ,3 ]
Sedykh, Anna V. [1 ]
Vetrova, Irina I. [2 ]
Sergienko, Elena A. [2 ]
机构
[1] Lomonosov Moscow State Univ, Moscow, Russia
[2] Russian Acad Sci, Inst Psychol, Moscow, Russia
[3] Fed Sci Ctr Psychol & Interdisciplinary Res, Moscow, Russia
来源
PSYCHOLOGY IN RUSSIA-STATE OF THE ART | 2024年 / 17卷 / 02期
关键词
artificial empathy; artificial psychology; ChatGPT; emotional intelligence (EI); emotional quotient(EQ); GPT-4; machine behavior;
D O I
10.11621/pir.2024.0206
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Background. Advanced AI models such as the large language model GPT-4 demonstrate sophisticated intellectual capabilities, sometimes exceeding human intellectual performance. However, the emotional competency of these models, along with their underlying mechanisms, has not been sufficiently evaluated. Objective. Our research aimed to explore different emotional intelligence domains in GPT-4 according to the Mayer-Salovey-Caruso model. We also tried to find out whether GPT-4's answer accuracy is consistent with its explanation of the answer. Design. The Russian version of the Mayer-Salovey-Caruso Emotional Intelligence Test (MSCEIT) sections was used in this research, with questions asked as text prompts in separate, independent ChatGPT chats three times each. Results. . High scores were achieved by the GPT-4 Large Language Model on the Understanding Emotions scale (with scores of 117, 124, and 128 across the three runs) and the Strategic Emotional Intelligence scale (with scores of 118, 121, and 122). Average scores were obtained on the Managing Emotions scale (103, 108, and 110 points). However, the Using Emotions to Facilitate Thought scale yielded low and less reliable scores (85, 86, and 88 points). Four types of explanations for the answer choices were identified: Meaningless sentences; Relation declaration; Implicit logic; and Explicit logic. Correct answers were accompanied by all types of explanations, whereas incorrect answers were only followed by Meaningless sentences or Explicit logic. This distribution aligns with observed patterns in children when they explore and elucidate mental states. Conclusion. GPT-4 is capable of emotion identification and managing emotions, but it lacks deep reflexive analysis of emotional experience and the motivational aspect of emotions.
引用
收藏
页码:85 / 99
页数:15
相关论文
共 50 条
  • [21] Evaluating Large Language Model-Assisted Emergency Triage: A Comparison of Acuity Assessments by GPT-4 and Medical Experts
    Haim, Gal Ben
    Saban, Mor
    Barash, Yiftach
    Cirulnik, David
    Shaham, Amit
    Eisenman, Ben Zion
    Burshtein, Livnat
    Mymon, Orly
    Klang, Eyal
    JOURNAL OF CLINICAL NURSING, 2024,
  • [22] Investigating the clinical reasoning abilities of large language model GPT-4: an analysis of postoperative complications from renal surgeries
    Hsueh, Jessica Y.
    Nethala, Daniel
    Singh, Shiva
    Linehan, W. Marston
    Ball, Mark W.
    UROLOGIC ONCOLOGY-SEMINARS AND ORIGINAL INVESTIGATIONS, 2024, 42 (09) : 292e1 - 292e7
  • [23] ChatGPT/GPT-4 (large language models): Opportunities and challenges of perspective in bariatric healthcare professionals
    Law, Saikam
    Oldfield, Brian
    Yang, Wah
    OBESITY REVIEWS, 2024, 25 (07)
  • [24] Evaluating Large Language Models for the National Premedical Exam in India: Comparative Analysis of GPT-3.5, GPT-4, and Bard
    Farhat, Faiza
    Chaudhry, Beenish Moalla
    Nadeem, Mohammad
    Sohail, Shahab Saquib
    Madsen, Dag Oivind
    JMIR MEDICAL EDUCATION, 2024, 10
  • [25] ARTIFICIAL REASON AND ARTIFICIAL INTELLIGENCE: THE LEGAL REASONING CAPABILITIES OF GPT-4
    Spaic, Bojan
    Jovanovic, Miodrag
    ANNALS OF THE FACULTY OF LAW IN BELGRADE, 2024, 72 (03): : 383 - 422
  • [26] Is GPT-4 a reliable rater? Evaluating consistency in GPT-4's text ratings
    Hackl, Veronika
    Mueller, Alexandra Elena
    Granitzer, Michael
    Sailer, Maximilian
    FRONTIERS IN EDUCATION, 2023, 8
  • [27] Stratified Evaluation of Large Language Model GPT-4's Question-Answering In Surgery reveals AI Knowledge Gaps
    Lonergan, Rebecca Murphy
    Curry, Jake
    Dhas, Kallpana
    Simmons, Benno
    BRITISH JOURNAL OF SURGERY, 2024, 111
  • [28] GPT-4 as a biomedical simulator
    Schaefer M.
    Reichl S.
    ter Horst R.
    Nicolas A.M.
    Krausgruber T.
    Piras F.
    Stepper P.
    Bock C.
    Samwald M.
    Computers in Biology and Medicine, 2024, 178
  • [29] Fine-Tuning Large Language Models for Ontology Engineering: A Comparative Analysis of GPT-4 and Mistral
    Doumanas, Dimitrios
    Soularidis, Andreas
    Spiliotopoulos, Dimitris
    Vassilakis, Costas
    Kotis, Konstantinos
    APPLIED SCIENCES-BASEL, 2025, 15 (04):
  • [30] Evaluation Metrics in the Era of GPT-4: Reliably Evaluating Large Language Models on Sequence to Sequence Tasks
    Sottana, Andrea
    Liang, Bin
    Zou, Kai
    Yuan, Zheng
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 8776 - 8788