Fear emotion classification in speech by acoustic and behavioral cues

被引:6
|
作者
Yoon, Shin-ae [1 ]
Son, Guiyoung [1 ]
Kwon, Soonil [1 ]
机构
[1] Sejong Univ, Coll Software & Convergence Technol, Dept Software, 209 Neung Dong Ro, Seoul 05006, South Korea
关键词
Emotional speech classification; Emergency situation; Behavioral cue; Disfluency(interjection); Speech signal processing; VOCAL EXPRESSION; RECOGNITION; COMMUNICATION; DISCRETE; FEATURES;
D O I
10.1007/s11042-018-6329-2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine-based emotional speech classification has become a requirement for natural and familiar human-computer interactions. Because emotional speech recognition systems use a person's voice to spontaneously detect their emotional status and take subsequent appropriate actions, they can be used widely for various reason in call centers and emotional based media services. Emotional speech recognition systems are primarily developed using emotional acoustic data. While there are several emotional acoustic databases available for emotion recognition systems in other countries, there is currently no real situational data related to the fear emotion available. Thus, in this study, we collected acoustic data recordings which represent real urgent and fearful situations from an emergency call center. To classify callers' emotions more accurately, we also included the additional behavioral feature interjection which can be classified as a type of disfluency which arises due to cognitive dysfunction observed in spontaneous speech when a speaker gets hyperemotional. We used Support Vector Machines (SVM), with the interjections feature, as well as conventionally used acoustic features (i.e., F0 variability, voice intensity variability, and Mel-Frequency Cepstral Coefficients; MFCCs) to identify which emotional category acoustic data fell into. The results of our study revealed that the MFCC was the best acoustic feature for spontaneous fear speech classification. In addition, we demonstrated the validity of behavioral features as an important criteria for emotional classification improvement.
引用
收藏
页码:2345 / 2366
页数:22
相关论文
共 50 条
  • [41] Investigating Acoustic Cues in Automatic Detection of Learners' Emotion from Auto Tutor
    Sun, Rui
    Moore, Elliot, II
    AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION, PT II, 2011, 6975 : 91 - 100
  • [42] Detecting emotion in speech expressing incongruent emotional cues through voice and content: investigation on dominant modality and language
    Kikutani, Mariko
    Ikemoto, Machiko
    COGNITION & EMOTION, 2022, 36 (03) : 492 - 511
  • [43] Filters Know How You Feel: Explaining Intermediate Speech Emotion Classification Representations
    Anand, Anubhav
    Negi, Shubham
    Narendra, N.
    2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 756 - 760
  • [44] Emotion Classification with EEG Responses Evoked by Emotional Prosody of Speech
    Zhang, Zechen
    Wu, Xihong
    Chen, Jing
    INTERSPEECH 2023, 2023, : 4254 - 4258
  • [46] Speech Emotion Classification Using Attention-Based LSTM
    Xie, Yue
    Liang, Ruiyu
    Liang, Zhenlin
    Huang, Chengwei
    Zou, Cairong
    Schuller, Bjoern
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2019, 27 (11) : 1675 - 1685
  • [47] Survey of deep emotion recognition in dynamic data using facial, speech and textual cues
    Zhang, Tao
    Tan, Zhenhua
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (25) : 66223 - 66262
  • [48] Evaluation of Speech Emotion Classification Based on GMM and Data Fusion
    Vondra, Martin
    Vich, Robert
    CROSS-MODAL ANALYSIS OF SPEECH, GESTURES, GAZE AND FACIAL EXPRESSIONS, 2009, 5641 : 98 - 105
  • [49] Music Theory-Inspired Acoustic Representation for Speech Emotion Recognition
    Li, Xingfeng
    Shi, Xiaohan
    Hu, Desheng
    Li, Yongwei
    Zhang, Qingchen
    Wang, Zhengxia
    Unoki, Masashi
    Akagi, Masato
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2023, 31 : 2534 - 2547
  • [50] The Impact of Clear and Loud Speech Cues on Acoustic and Perceptual Features of Speech Production in Adolescents With Down Syndrome
    Darling-White, Meghan
    Mchugh, Allison
    AMERICAN JOURNAL OF SPEECH-LANGUAGE PATHOLOGY, 2025, 34 (01) : 201 - 217