Enhancement of emotion detection in spoken dialogue systems by combining several information sources

被引:13
|
作者
Lopez-Cozar, Ramon [1 ]
Silovsky, Jan [2 ]
Kroul, Martin [2 ]
机构
[1] Univ Granada, Dept Languages & Comp Syst, Fac Comp Sci, E-18071 Granada, Spain
[2] Tech Univ Liberec, Inst Informat Technol & Elect, Fac Mechatron, Liberec, Czech Republic
关键词
Adaptive spoken dialogue systems; Combination of classifiers; Information fusion; Emotion detection; Human computer interaction; RECOGNITION; AGREEMENT; USER;
D O I
10.1016/j.specom.2011.01.006
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper proposes a technique to enhance emotion detection in spoken dialogue systems by means of two modules that combine different information sources. The first one, called Fusion-0, combines emotion predictions generated by a set of classifiers that deal with different kinds of information about each sentence uttered by the user. To do this, the module employs several methods for information fusion that produce other predictions about the emotional state of the user. The predictions are the input to the second information fusion module, called Fusion-1, where they are combined to deduce the emotional state of the user. Fusion-0 represents a method employed in previous studies to enhance classification rates, whereas Fusion-1 represents the novelty of the technique, which is the combination of emotion predictions generated by Fusion-0. One advantage of the technique is that it can be applied as a posterior processing stage to any other methods that combine information from different information sources at the decision level. This is so because the technique works on the predictions (outputs) of the methods, without interfering in the procedure used to obtain these predictions. Another advantage is that the technique can be implemented as a modular architecture, which facilitates the setting up within a spoken dialogue system as well as the deduction of the emotional state of the user in real time. Experiments have been carried out considering classifiers to deal with prosodic, acoustic, lexical, and dialogue acts information, and three methods to combine information: multiplication of probabilities, average of probabilities, and unweighted vote. The results show that the technique enhances the classification rates of the standard fusion by 2.27% and 3.38% absolute in experiments carried out considering two and three emotion categories, respectively. (C) 2011 Elsevier B.V. All rights reserved.
引用
收藏
页码:1210 / 1228
页数:19
相关论文
共 44 条
  • [1] Enhancement of Spoken Dialogue Systems by Means of User Emotion Recognition
    Lopez-Cozar, Ramon
    Silovsky, Jan
    Griol, David
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2010, (45): : 191 - 198
  • [2] Influence of contextual information in emotion annotation for spoken dialogue systems
    Callejas, Zoraida
    Lopez-Cozar, Ramon
    SPEECH COMMUNICATION, 2008, 50 (05) : 416 - 433
  • [3] Emotion recognition and adaptation in spoken dialogue systems
    Pittermann, Johannes
    Pittermann, Angela
    Minker, Wolfgang
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2010, 13 (01) : 49 - 60
  • [4] Predicting emotion in spoken dialogue from multiple knowledge sources
    Forbes-Riley, K
    Litman, DJ
    HLT-NAACL 2004: HUMAN LANGUAGE TECHNOLOGY CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE MAIN CONFERENCE, 2004, : 201 - 208
  • [5] Optimising Information Presentation for Spoken Dialogue Systems
    Rieser, Verena
    Lemon, Oliver
    Liu, Xingkun
    ACL 2010: 48TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2010, : 1009 - 1018
  • [6] Information seeking spoken dialogue systems - Part II: Multimodal dialogue
    Potamianos, Alexandros
    Fosler-Lussier, Eric
    Ammicht, Egbert
    Perakakis, Manolis
    IEEE TRANSACTIONS ON MULTIMEDIA, 2007, 9 (03) : 550 - 566
  • [7] OUT-OF-DOMAIN SLOT VALUE DETECTION FOR SPOKEN DIALOGUE SYSTEMS WITH CONTEXT INFORMATION
    Kobayashi, Yuka
    Yoshida, Takami
    Iwata, Kenji
    Fujimura, Hiroshi
    Akamine, Masami
    2018 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2018), 2018, : 854 - 861
  • [8] Knowledge-Combining Methodology for Dialogue Design in Spoken Language Systems
    Rubén San-Segundo
    Juan M. Montero
    Javier Macías-Guarasa
    Javier Ferreiros
    José M. Pardo
    International Journal of Speech Technology, 2005, 8 (1) : 45 - 66
  • [9] Knowledge-Combining Methodology for Dialogue Design in Spoken Language Systems
    San-Segundo, Ruben
    Montero, Juan M.
    Macias-Guarasa, Javier
    Ferreiros, Javier
    Pardo, Jose M.
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2005, 8 (01) : 45 - 66
  • [10] Recognition of Paralinguistic Information in Spoken Dialogue Systems for Elderly People
    Perez-Espinosa, Humberto
    Martinez-Miranda, Juan
    ADVANCES IN ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING, MICAI 2015, PT I, 2015, 9413 : 107 - 117