Annotating and modeling empathy in spoken conversations

被引:47
作者
Alam, Firoj [1 ]
Danieli, Morena [1 ]
Riccardi, Giuseppe [1 ]
机构
[1] Univ Trento, Dept Informat Engn & Comp Sci, I-38123 Trento, Italy
关键词
Empathy; Emotion; Spoken conversation; Behavior analysis; Affective scene; Affect; Call center; Human-Human conversation; CLASSIFICATION; SENTIMENT; DIFFUSION; AGREEMENT; EMOTIONS; PROSODY; SPEECH; WORDS; MOOD;
D O I
10.1016/j.csl.2017.12.003
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Empathy, as defined in behavioral sciences, expresses the ability of human beings to recognize, understand and react to emotions, attitudes and beliefs of others. In this paper, we address two related problems in automatic affective behavior analysis: the design of the annotation protocol and the automatic recognition of empathy from human human dyadic spoken conversations. We propose and evaluate an annotation scheme for empathy inspired by the modal model of emotions. The annotation scheme was evaluated on a corpus of real-life, dyadic spoken conversations. In the context of behavioral analysis, we designed an automatic segmentation and classification system for empathy. Given the different speech and language levels of representation where empathy may be communicated, we investigated features derived from the lexical and acoustic spaces. The feature development process was designed to support both the fusion and automatic selection of relevant features from a high dimensional space. The automatic classification system was evaluated on call center conversations where it showed significantly better performance than the baseline. (C) 2017 Elsevier Ltd. All rights reserved.
引用
收藏
页码:40 / 61
页数:22
相关论文
共 95 条
[61]  
Kumano S., 2011, Proceedings 2011 IEEE International Conference on Automatic Face & Gesture Recognition (FG 2011), P43, DOI 10.1109/FG.2011.5771440
[62]   MEASUREMENT OF OBSERVER AGREEMENT FOR CATEGORICAL DATA [J].
LANDIS, JR ;
KOCH, GG .
BIOMETRICS, 1977, 33 (01) :159-174
[63]  
Lerch Alexander, 2012, INTRO AUDIO CONTENT
[64]  
Lin C., 2009, P 18 ACM C INF KNOWL, P375, DOI [10.1145/1645953.1646003, DOI 10.1145/1645953.1646003]
[65]  
Liscombe J., 2005, Interspeech, P1845
[66]   The SEMAINE Database: Annotated Multimodal Records of Emotionally Colored Conversations between a Person and a Limited Agent [J].
McKeown, Gary ;
Valstar, Michel ;
Cowie, Roddy ;
Pantic, Maja ;
Schroeder, Marc .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2012, 3 (01) :5-17
[67]   NOTE ON THE SAMPLING ERROR OF THE DIFFERENCE BETWEEN CORRELATED PROPORTIONS OR PERCENTAGES [J].
McNemar, Quinn .
PSYCHOMETRIKA, 1947, 12 (02) :153-157
[68]  
Metallinou A., 2013, PROCEEDINGSOF10THIEE, P1
[69]   Context-Sensitive Learning for Enhanced Audiovisual Emotion Classification [J].
Metallinou, Angeliki ;
Woellmer, Martin ;
Katsamanis, Athanasios ;
Eyben, Florian ;
Schuller, Bjoern ;
Narayanan, Shrikanth .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2012, 3 (02) :184-198
[70]  
Nasukawa T., 2003, K CAP, P70, DOI DOI 10.1145/945645.945658