Deep Learning, Ensemble and Supervised Machine Learning for Arabic Speech Emotion Recognition

被引:0
|
作者
Ismaiel, Wahiba
Alhalangy, Abdalilah [1 ,2 ]
Mohamed, Adil. O. Y. [2 ]
Musa, Abdalla Ibrahim Abdalla [2 ]
机构
[1] Taif Univ, Univ Coll Ranyah, Dept Sci & Technol, Taif, Saudi Arabia
[2] Qassim Univ, Coll Comp, Dept Comp Engn, Buraydah, Saudi Arabia
关键词
Arabic speech emotion recognition; ANAD; SERDNN; SOM; Xgboost; Adaboost; DT; KNN; RANDOM FOREST;
D O I
10.48084/etasr.7134
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Today, automatic emotion recognition in speech is one of the most important areas of research in signal processing. Identifying emotional content in Arabic speech is regarded as a very challenging and intricate task due to several obstacles, such as the wide range of cultures and dialects, the influence of cultural factors on emotional expression, and the scarcity of available datasets. This study used a variety of artificial intelligence models, including Xgboost, Adaboost, KNN, DT, and SOM, and a deep -learning model named SERDNN. ANAD was employed as a training dataset, which contains three emotions, "angry", "happy", and "surprised", with 844 features. This study aimed to present a more efficient and accurate technique for recognizing emotions in Arabic speech. Precision, accuracy, recall, and F1 -score metrics were utilized to evaluate the effectiveness of the proposed techniques. The results showed that the Xgboost, SOM, and KNN classifiers achieved superior performance in recognizing emotions in Arabic speech. The SERDNN deep learning model outperformed the other techniques, achieving the highest accuracy of 97.40% with a loss rate of 0.1457. Therefore, it can be relied upon and deployed to recognize emotions in Arabic speech.
引用
收藏
页码:13757 / 13764
页数:8
相关论文
共 50 条
  • [1] Ensemble deep learning with HuBERT for speech emotion recognition
    Yang, Janghoon
    2023 IEEE 17TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING, ICSC, 2023, : 153 - 154
  • [2] SPEECH EMOTION RECOGNITION WITH ENSEMBLE LEARNING METHODS
    Shih, Po-Yuan
    Chen, Chia-Ping
    Wu, Chung-Hsien
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 2756 - 2760
  • [3] Speech Emotion Recognition with Deep Learning
    Harar, Pavol
    Burget, Radim
    Dutta, Malay Kishore
    2017 4TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND INTEGRATED NETWORKS (SPIN), 2017, : 137 - 140
  • [4] A computationally efficient speech emotion recognition system employing machine learning classifiers and ensemble learning
    Aishwarya N.
    Kaur K.
    Seemakurthy K.
    International Journal of Speech Technology, 2024, 27 (1) : 239 - 254
  • [5] Speech emotion recognition for psychotherapy: an analysis of traditional machine learning and deep learning techniques
    Shah, Nidhi
    Sood, Kanika
    Arora, Jayraj
    2023 IEEE 13TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE, CCWC, 2023, : 718 - 723
  • [6] Arabic Speech Recognition with Deep Learning: A Review
    Algihab, Wajdan
    Alawwad, Noura
    Aldawish, Anfal
    AlHumoud, Sarah
    SOCIAL COMPUTING AND SOCIAL MEDIA: DESIGN, HUMAN BEHAVIOR AND ANALYTICS, SCSM 2019, PT I, 2019, 11578 : 15 - 31
  • [7] Emotion Recognition on Multimodal with Deep Learning and Ensemble
    Dharma, David Adi
    Zahra, Amalia
    International Journal of Advanced Computer Science and Applications, 2022, 13 (12): : 656 - 663
  • [8] Emotion Recognition on Multimodal with Deep Learning and Ensemble
    Dharma, David Adi
    Zahra, Amalia
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (12) : 656 - 663
  • [9] Multiview Supervised Dictionary Learning in Speech Emotion Recognition
    Gangeh, Mehrdad J.
    Fewzee, Pouria
    Ghodsi, Ali
    Kamel, Mohamed S.
    Karray, Fakhri
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2014, 22 (06) : 1056 - 1068
  • [10] Machine Learning Approach for Emotion Recognition in Speech
    Gjoreski, Martin
    Gjoreski, Hristijan
    Kulakov, Andrea
    INFORMATICA-JOURNAL OF COMPUTING AND INFORMATICS, 2014, 38 (04): : 377 - 383