Speech Emotion Recognition using Feature Selection with Adaptive Structure Learning

被引:1
作者
Rayaluru, Akshay [1 ]
Bandela, Surekha Reddy [1 ]
Kumar, T. Kishore [1 ]
机构
[1] NIT Warangal, Dept ECE, Fathimanagar, Telangana, India
来源
2019 IEEE INTERNATIONAL SYMPOSIUM ON SMART ELECTRONIC SYSTEMS (ISES 2019) | 2019年
关键词
Speech Emotion Recognition; INTERSPEECH Paralinguistic Feature Set; Feature Selection; FSASL;
D O I
10.1109/iSES47678.2019.00059
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Speech Emotion Recognition(SER) has gained a lot of interest in recent times. The combination of different speech features improves the accuracy of the SER system. Whereas, this results in an increase of the time taken by the classifier to train the huge feature set. Also, there are some of the features that could not be useful for emotion recognition which leads to the decrease in the recognition accuracy. Therefore, in order to surmount this disadvantage, feature selection algorithms can be used in order to choose the most prominent features that could contribute highly for classification of the emotions efficiently. In this paper, a Feature Selection with Adaptive Structure Learning (FSASL) is used for selecting the appropriate features for SER. In the proposed SER system, the 1582 INTERSPEECH 2010 Paralinguistic features are extracted from the speech signal and the FSASL Feature Selection algorithm is used for selecting the best features from the huge feature set. The SVM and k-NN classifiers with 5-fold cross-validation scheme is used for classifying the emotions. EMO-DB, Berlin German database is used in this work and the Classification accuracy performance metric are considered for the evaluation of the proposed SER system. The results emphasize that the classification accuracy of the proposed SER system is improved remarkably upon using the FSASL algorithm as compared to the baseline as well as the existing SER systems.
引用
收藏
页码:233 / 236
页数:4
相关论文
共 16 条
[1]  
Chen S.-H., 2016, ALTERNATIVE MED, P1, DOI DOI 10.1155/2016/7137362
[2]   Multiscale Amplitude Feature and Significance of Enhanced Vocal Tract Information for Emotion Classification [J].
Deb, Suman ;
Dandapat, Samarendra .
IEEE TRANSACTIONS ON CYBERNETICS, 2019, 49 (03) :802-815
[3]  
Du Liang, P 21 ACM SIGKDD INT, P209
[4]   Survey on speech emotion recognition: Features, classification schemes, and databases [J].
El Ayadi, Moataz ;
Kamel, Mohamed S. ;
Karray, Fakhri .
PATTERN RECOGNITION, 2011, 44 (03) :572-587
[5]  
Eyben Florian, P 18TH ACM INT C
[6]   Emotion recognition from speech using sub-syllabic and pitch synchronous spectral features [J].
Koolagudi, Shashidhar ;
Krothapalli, Sreenivasa .
INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2012, 15 (04) :495-511
[7]   An optimal two stage feature selection for speech emotion recognition using acoustic features [J].
Kuchibhotla S. ;
Vankayalapati H.D. ;
Anne K.R. .
International Journal of Speech Technology, 2016, 19 (04) :657-667
[8]   Speech emotion recognition based on an improved brain emotion learning model [J].
Liu, Zhen-Tao ;
Xie, Qiao ;
Wu, Min ;
Cao, Wei-Hua ;
Mei, Ying ;
Mao, Jun-Wei .
NEUROCOMPUTING, 2018, 309 :145-156
[9]  
Lugger M, 2007, INT CONF ACOUST SPEE, P17
[10]  
Schuller B, 2010, 11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 3 AND 4, P2798