Deep ensemble models for speech emotion classification

被引:4
|
作者
Pravin, Sheena Christabel [1 ]
Sivaraman, Vishal Balaji [2 ]
Saranya, J. [3 ]
机构
[1] Vellore Inst Technol, Sch Elect Engn SENSE, Chennai, India
[2] Univ Florida, Gainesville, FL USA
[3] Rajalakshmi Engn Coll, Thandalam, India
关键词
Deep cascaded ensemble; Deep parallel ensemble; Speech emotion classification; Memory consumption and run time complexity; RECOGNITION;
D O I
10.1016/j.micpro.2023.104790
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
This research article proposes two deep ensemble models, namely the Deep Cascaded Ensemble (DCE) and Deep Parallel Ensemble (DPE) for automatic speech emotion classification. Classification of emotions into their respective classes has so long relied on machine learning and deep learning networks. The proposed models are a blend of different machine learning and deep learning models in a cascaded/parallel architecture. The proposed models have exhibited a considerable reduction in the consumption of memory in a Google Colab environment. Furthermore, the issues of tuning numerous hyper-parameters and the huge data demand of the deep learning algorithms are overcome by the proposed deep ensemble models. The proposed DCE and DPE have yielded optimal classification accuracy with reduced consumption of memory over less data. Experimentally, the pro-posed deep cascaded ensemble has superior performance compared to the deep parallel ensemble of the same combination of deep learning and machine learning networks as well. The proposed models and the baseline models were evaluated in terms of possible performance metrics including Cohen's kappa coefficient, accuracy of classification accuracy, space and time complexity.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] A New Amharic Speech Emotion Dataset and Classification Benchmark
    Retta, Ephrem Afele
    Almekhlafi, Eiad
    Sutcliffe, Richard
    Mhamed, Mustafa
    Ali, Haider
    Feng, Jun
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (01)
  • [22] Fear emotion classification in speech by acoustic and behavioral cues
    Yoon, Shin-ae
    Son, Guiyoung
    Kwon, Soonil
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (02) : 2345 - 2366
  • [23] Paralinguistic and spectral feature extraction for speech emotion classification using machine learning techniques
    Liu, Tong
    Yuan, Xiaochen
    EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2023, 2023 (01)
  • [24] Automatic Speech Emotion Detection System using Multi-domain Acoustic Feature Selection and Classification Models
    Semwal, Nancy
    Kumar, Abhijeet
    Narayanan, Sakthivel
    2017 IEEE INTERNATIONAL CONFERENCE ON IDENTITY, SECURITY AND BEHAVIOR ANALYSIS (ISBA), 2017,
  • [25] Speech Emotion Classification Using Attention-Based LSTM
    Xie, Yue
    Liang, Ruiyu
    Liang, Zhenlin
    Huang, Chengwei
    Zou, Cairong
    Schuller, Bjoern
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2019, 27 (11) : 1675 - 1685
  • [26] Speech Based Emotion Classification Framework for Driver Assistance System
    Tawari, Ashish
    Trivedi, Mohan
    2010 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2010, : 174 - 178
  • [27] Emotion Classification with EEG Responses Evoked by Emotional Prosody of Speech
    Zhang, Zechen
    Wu, Xihong
    Chen, Jing
    INTERSPEECH 2023, 2023, : 4254 - 4258
  • [28] Evaluation of Speech Emotion Classification Based on GMM and Data Fusion
    Vondra, Martin
    Vich, Robert
    CROSS-MODAL ANALYSIS OF SPEECH, GESTURES, GAZE AND FACIAL EXPRESSIONS, 2009, 5641 : 98 - 105
  • [29] Speech Emotion Classification Using Multiple Kernel Gaussian Process
    Chen, Sih-Huei
    Wang, Jia-Ching
    Hsieh, Wen-Chi
    Chin, Yu-Hao
    Ho, Chin-Wen
    Wu, Chung-Hsien
    2016 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2016,
  • [30] Internal Emotion Classification Using EEG Signal With Sparse Discriminative Ensemble
    Ullah, Habib
    Uzair, Muhammad
    Mahmood, Arif
    Ullah, Mohib
    Khan, Sultan Daud
    Cheikh, Faouzi Alaya
    IEEE ACCESS, 2019, 7 : 40144 - 40153