Fear emotion classification in speech by acoustic and behavioral cues

被引:6
|
作者
Yoon, Shin-ae [1 ]
Son, Guiyoung [1 ]
Kwon, Soonil [1 ]
机构
[1] Sejong Univ, Coll Software & Convergence Technol, Dept Software, 209 Neung Dong Ro, Seoul 05006, South Korea
关键词
Emotional speech classification; Emergency situation; Behavioral cue; Disfluency(interjection); Speech signal processing; VOCAL EXPRESSION; RECOGNITION; COMMUNICATION; DISCRETE; FEATURES;
D O I
10.1007/s11042-018-6329-2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine-based emotional speech classification has become a requirement for natural and familiar human-computer interactions. Because emotional speech recognition systems use a person's voice to spontaneously detect their emotional status and take subsequent appropriate actions, they can be used widely for various reason in call centers and emotional based media services. Emotional speech recognition systems are primarily developed using emotional acoustic data. While there are several emotional acoustic databases available for emotion recognition systems in other countries, there is currently no real situational data related to the fear emotion available. Thus, in this study, we collected acoustic data recordings which represent real urgent and fearful situations from an emergency call center. To classify callers' emotions more accurately, we also included the additional behavioral feature interjection which can be classified as a type of disfluency which arises due to cognitive dysfunction observed in spontaneous speech when a speaker gets hyperemotional. We used Support Vector Machines (SVM), with the interjections feature, as well as conventionally used acoustic features (i.e., F0 variability, voice intensity variability, and Mel-Frequency Cepstral Coefficients; MFCCs) to identify which emotional category acoustic data fell into. The results of our study revealed that the MFCC was the best acoustic feature for spontaneous fear speech classification. In addition, we demonstrated the validity of behavioral features as an important criteria for emotional classification improvement.
引用
收藏
页码:2345 / 2366
页数:22
相关论文
共 50 条
  • [1] Fear emotion classification in speech by acoustic and behavioral cues
    Shin-ae Yoon
    Guiyoung Son
    Soonil Kwon
    Multimedia Tools and Applications, 2019, 78 : 2345 - 2366
  • [2] Psychoacoustic cues to emotion in speech prosody and music
    Coutinho, Eduardo
    Dibben, Nicola
    COGNITION & EMOTION, 2013, 27 (04) : 658 - 684
  • [3] Common cues to emotion in the dynamic facial expressions of speech and song
    Livingstone, Steven R.
    Thompson, William F.
    Wanderley, Marcelo M.
    Palmer, Caroline
    QUARTERLY JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 2015, 68 (05) : 952 - 970
  • [4] Investigating Acoustic Cues of Emotional Valence in Mandarin Speech Prosody - A Corpus Approach
    Li, Junlin
    Huang, Chu-Ren
    CHINESE LEXICAL SEMANTICS, CLSW 2023, PT II, 2024, 14515 : 316 - 330
  • [5] Measuring the Randomness of Speech Cues for Emotion Recognition
    Susan, Seba
    Kaur, Amandeep
    2017 TENTH INTERNATIONAL CONFERENCE ON CONTEMPORARY COMPUTING (IC3), 2017, : 78 - 83
  • [6] Exploration of Phase Information for Speech Emotion Classification
    Deb, Suman
    Dandapat, S.
    2017 TWENTY-THIRD NATIONAL CONFERENCE ON COMMUNICATIONS (NCC), 2017,
  • [7] Cultural dependency analysis for understanding speech emotion
    Kamaruddin, Norhaslinda
    Wahab, Abdul
    Quek, Chai
    EXPERT SYSTEMS WITH APPLICATIONS, 2012, 39 (05) : 5115 - 5133
  • [8] Feature Analysis for Speech Emotion Classification
    Kingeski, R.
    Schueda, L. A. P.
    Paterno, A. S.
    XXVII BRAZILIAN CONGRESS ON BIOMEDICAL ENGINEERING, CBEB 2020, 2022, : 2359 - 2365
  • [9] Preliminary Arabic Speech Emotion Classification
    Meftah, Ali
    Selouani, Sid-Ahmed
    Alotaibi, Yousef A.
    2014 IEEE INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING AND INFORMATION TECHNOLOGY (ISSPIT), 2014, : 179 - 182
  • [10] Speech Based Emotion Classification Framework for Driver Assistance System
    Tawari, Ashish
    Trivedi, Mohan
    2010 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2010, : 174 - 178