Recognizing Semi-Natural and Spontaneous Speech Emotions Using Deep Neural Networks

被引:8
作者
Amjad, Ammar [1 ]
Khan, Lal [1 ]
Ashraf, Noman [2 ]
Mahmood, Muhammad Bilal [3 ]
Chang, Hsien-Tsung [1 ,4 ,5 ,6 ]
机构
[1] Chang Gung Univ, Coll Engn, Dept Comp Sci & Informat Engn, Taoyuan 33302, Taiwan
[2] Inst Politecn Nacl, CIC, Mexico City 7738, DF, Mexico
[3] Dalian Univ Technol, Dept Software Engn, Dalian 116024, Peoples R China
[4] Chang Gung Univ, Artificial Intelligence Res Ctr, Taoyuan 33302, Taiwan
[5] Chang Gung Mem Hosp, Dept Phys Med & Rehabil, Taoyuan 333, Taiwan
[6] Chang Gung Univ, Bachelor Program Artificial Intelligence, Taoyuan 33302, Taiwan
来源
IEEE ACCESS | 2022年 / 10卷
关键词
Feature extraction; Convolutional neural networks; Spectrogram; Databases; Emotion recognition; Speech recognition; Data models; Speech emotion recognition; convolutional neural network; data augmentation; long-short-term memory; spontaneous speech database; RECOGNITION FEATURES; SENTIMENT ANALYSIS; CLASSIFICATION;
D O I
10.1109/ACCESS.2022.3163712
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We needed to find deep emotional features to identify emotions from audio signals. Identifying emotions in spontaneous speech is a novel and challenging subject of research. Several convolutional neural network (CNN) models were used to learn deep segment-level auditory representations of augmented Mel spectrograms. The proposed study introduces a novel technique for recognizing semi-natural and spontaneous speech emotions based on 1D (Model A) and 2D (Model B) deep convolutional neural networks (DCNNs) with two layers of long-short-term memory (LSTM). Both models used raw speech data and augmented (mid, left, right, and side) segment level Mel spectrograms to learn local and global features. The architecture of both models consists of five local feature learning blocks (LFLBs), two LSTM layers, and a fully connected layer (FCL). In addition to learning local correlations and extracting hierarchical correlations, LFLB comprises two convolutional layers and a max-pooling layer. The LSTM layer learns long-term correlations from local features. The experiments illustrated that the proposed systems perform better than conventional methods. Model A achieved an average identification accuracy of 94.78% for speaker-dependent (SD) with a raw SAVEE dataset. With the IEMOCAP database, Model A achieved an average accuracy of an SD experiment with raw audio of 73.15%. In addition, Model A obtained identification accuracies of 97.19%, 94.09%, and 53.98% on SAVEE, IEMOCAP, and BAUM-1s, the databases for speaker-dependent (SD) experiments with an augmented Mel spectrogram, respectively. In contrast, Model B achieved identification accuracy of 96.85%, 88.80%, and 48.67% on SAVEE, IEMOCAP, and the BAUM-1s database for SI experiments with augmented reality Mel spectrogram, respectively.
引用
收藏
页码:37149 / 37163
页数:15
相关论文
共 95 条
  • [1] Deep Learning Techniques for Speech Emotion Recognition, from Databases to Models
    Abbaschian, Babak Joze
    Sierra-Sosa, Daniel
    Elmaghraby, Adel
    [J]. SENSORS, 2021, 21 (04) : 1 - 27
  • [2] Phonemic segmentation using the generalised Gamma distribution and small sample Bayesian information criterion
    Almpanidis, George
    Kotropoulos, Constantine
    [J]. SPEECH COMMUNICATION, 2008, 50 (01) : 38 - 55
  • [3] Ameer I, 2020, COMPUT SIST, V24, P1159, DOI [10.13053/cys-24-3-3476, 10.13053/CyS-24-3-3476]
  • [4] Semi-Natural and Spontaneous Speech Recognition Using Deep Neural Networks with Hybrid Features Unification
    Amjad, Ammar
    Khan, Lal
    Chang, Hsien-Tsung
    [J]. PROCESSES, 2021, 9 (12)
  • [5] Effect on speech emotion classification of a feature selection approach using a convolutional neural network
    Amjad, Ammar
    Khan, Lal
    Chang, Hsien-Tsung
    [J]. PEERJ COMPUTER SCIENCE, 2021, 7
  • [6] [Anonymous], 2016, GITHUB KERAS TEAM KE
  • [7] Deep-Net: A Lightweight CNN-Based Speech Emotion Recognition System Using Deep Frequency Features
    Anvarjon, Tursunov
    Mustaqeem
    Kwon, Soonil
    [J]. SENSORS, 2020, 20 (18) : 1 - 16
  • [8] Ashraf M, 2018, INT J ADV COMPUT SC, V9, P427
  • [9] Abusive language detection in youtube comments leveraging replies as conversational context
    Ashraf, Noman
    Zubiaga, Arkaitz
    Gelbukh, Alexander
    [J]. PEERJ COMPUTER SCIENCE, 2021, 7 (07)
  • [10] Individual vs. Group Violent Threats Classification in Online Discussions
    Ashraf, Noman
    Mustafa, Rabia
    Sidorov, Grigori
    Gelbukh, Alexander
    [J]. WWW'20: COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2020, 2020, : 629 - 633