A modified feature selection method based on metaheuristic algorithms for speech emotion recognition

被引:65
作者
Yildirim, Serdar [1 ]
Kaya, Yasin [1 ]
Kilic, Fatih [1 ]
机构
[1] Adana Alparslan Turkes Sci & Technol Univ, Comp Engn Dept, Adana, Turkey
关键词
Feature selection; Emotion recognition from speech; Metaheuristic search algorithms; CUCKOO SEARCH ALGORITHM; OPTIMIZATION; CLASSIFICATION; DATABASES; NETWORK;
D O I
10.1016/j.apacoust.2020.107721
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Feature selection plays an important role to build a successful speech emotion recognition system. In this paper, a feature selection approach which modifies the initial population generation stage of metaheuristic search algorithms, is proposed. The approach is evaluated on two metaheuristic search algorithms, a nondominated sorting genetic algorithm-II (NSGA-II) and Cuckoo Search in the context of speech emotion recognition using Berlin emotional speech database (EMO-DB) and Interactive Emotional Dyadic Motion Capture (IEMOCAP) database. Results show that the presented feature selection algorithms reduce the number of features significantly and are still effective for emotion classification from speech. Specifically, in speaker-dependent experiments of the EMO-DB, recognition rates of 87.66% and 87.20% are obtained using selected features by modified Cuckoo Search and NSGA-II respectively, whereas, for the IEMOCAP database, the accuracies of 69.30% and 68.32% are obtained using SVM classifier. For the speaker-independent experiments, we achieved comparable results for both databases. Specifically, recognition rates of 76.80% and 76.82% for EMO-DB and 59.37% and 59.52% for IEMOCAP using modified NSGA-II and Cuckoo Search respectively. (C) 2020 Elsevier Ltd. All rights reserved.
引用
收藏
页数:11
相关论文
共 62 条
[1]   Modified cuckoo search algorithm with rough sets for feature selection [J].
Abd El Aziz, Mohamed ;
Hassanien, Aboul Ella .
NEURAL COMPUTING & APPLICATIONS, 2018, 29 (04) :925-934
[2]   Speech emotion recognition: Emotional models, databases, features, preprocessing methods, supporting modalities, and classifiers [J].
Akcay, Mehmet Berkehan ;
Oguz, Kaya .
SPEECH COMMUNICATION, 2020, 116 :56-76
[3]   Emotions Recognition Using EEG Signals: A Survey [J].
Alarcao, Soraia M. ;
Fonseca, Manuel J. .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2019, 10 (03) :374-393
[4]   Audiovisual emotion recognition in wild [J].
Avots, Egils ;
Sapinski, Tomasz ;
Bachmann, Maie ;
Kaminska, Dorota .
MACHINE VISION AND APPLICATIONS, 2019, 30 (05) :975-985
[5]   Deep features-based speech emotion recognition for smart affective services [J].
Badshah, Abdul Malik ;
Rahim, Nasir ;
Ullah, Noor ;
Ahmad, Jamil ;
Muhammad, Khan ;
Lee, Mi Young ;
Kwon, Soonil ;
Baik, Sung Wook .
MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (05) :5571-5589
[6]   Class-level spectral features for emotion recognition [J].
Bitouk, Dmitri ;
Verma, Ragini ;
Nenkova, Ani .
SPEECH COMMUNICATION, 2010, 52 (7-8) :613-625
[7]   MULTI-OBJECTIVE HEURISTIC FEATURE SELECTION FOR SPEECH-BASED MULTILINGUAL EMOTION RECOGNITION [J].
Brester, Christina ;
Semenkin, Eugene ;
Sidorov, Maxim .
JOURNAL OF ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING RESEARCH, 2016, 6 (04) :243-253
[8]  
Burkhardt F., 2005, Interspeech, P1517, DOI 10.21437/interspeech.2005-446
[9]  
Busso C., 2004, P 6 INT C MULT INT, P205, DOI [10.1145/1027933, DOI 10.1145/1027933.1027968]
[10]   IEMOCAP: interactive emotional dyadic motion capture database [J].
Busso, Carlos ;
Bulut, Murtaza ;
Lee, Chi-Chun ;
Kazemzadeh, Abe ;
Mower, Emily ;
Kim, Samuel ;
Chang, Jeannette N. ;
Lee, Sungbok ;
Narayanan, Shrikanth S. .
LANGUAGE RESOURCES AND EVALUATION, 2008, 42 (04) :335-359