Two-layer fuzzy multiple random forest for speech emotion recognition in human-robot interaction

被引:129
作者
Chen, Luefeng [1 ,2 ]
Su, Wanjuan [1 ,2 ]
Feng, Yu [1 ,2 ]
Wu, Min [1 ,2 ]
She, Jinhua [3 ]
Hirota, Kaoru [4 ,5 ]
机构
[1] China Univ Geosci, Sch Automat, Wuhan 430074, Hubei, Peoples R China
[2] Hubei Key Lab Adv Control & Intelligent Automat C, Wuhan 430074, Hubei, Peoples R China
[3] Tokyo Univ Technol, Sch Engn, Tokyo 1920982, Japan
[4] Tokyo Inst Technol, Yokohama, Kanagawa 2268502, Japan
[5] Beijing Inst Technol, Sch Automat, Beijing 100081, Peoples R China
基金
中国国家自然科学基金;
关键词
Speech emotion recognition; Fuzzy C-means; Multiple random forest; Human-robot interaction; CLASSIFICATION; REGRESSION; ALGORITHM; TIME;
D O I
10.1016/j.ins.2019.09.005
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The two-layer fuzzy multiple random forest (TLFMRF) is proposed for speech emotion recognition. When recognizing speech emotion, there are usually some problems. One is that feature extraction relies on personalized features. The other is that emotion recognition doesn't consider the differences among different categories of people. In the proposal, personalized and non-personalized features are fused for speech emotion recognition. High dimensional emotional features are divided into different subclasses by adopting the fuzzy C-means clustering algorithm, and multiple random forest is used to recognize different emotional states. Finally, a TLFMRF is established. Moreover, a separate classification of certain emotions which are difficult to recognize to some extent is conducted. The results show that the TLFMRF can identify emotions in a stable manner. To demonstrate the effectiveness of the proposal, experiments on CASIA corpus and Berlin EmoDB are conducted. Experimental results show the recognition accuracies of the proposal are 1.39%-7.64% and 4.06%-4.30% higher than that of back propagation neural network and random forest respectively. Meanwhile, preliminary application experiments are also conducted to investigate the emotional social robot system, and application results indicate that mobile robot can real-time track six basic emotions, including angry, fear, happy, neutral, sad, and surprise. (C) 2019 Elsevier Inc. All rights reserved.
引用
收藏
页码:150 / 163
页数:14
相关论文
共 50 条
[1]   Emotion Recognition in Never-Seen Languages Using a Novel Ensemble Method with Emotion Profiles [J].
Albornoz, E. M. ;
Milone, D. H. .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2017, 8 (01) :43-53
[2]   Feature extraction based on bio-inspired model for robust emotion recognition [J].
Albornoz, Enrique M. ;
Milone, Diego H. ;
Rufiner, Hugo L. .
SOFT COMPUTING, 2017, 21 (17) :5145-5158
[3]  
[Anonymous], 2008, CASIA CHINESE EMOTIO
[4]  
[Anonymous], 2005, BERLIN DATABASE EMOT
[5]  
BEZDEK JC, 1976, IEEE T SYST MAN CYB, V6, P387
[6]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[7]   The MatchNMingle Dataset: A Novel Multi-Sensor Resource for the Analysis of Social Interactions and Group Dynamics In-the-Wild During Free-Standing Conversations and Speed Dates [J].
Cabrera-Quiros, Laura ;
Demetriou, Andrew ;
Gedik, Ekin ;
van der Meij, Leander ;
Hung, Hayley .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2021, 12 (01) :113-130
[8]   Emotion-Age-Gender-Nationality Based Intention Understanding in Human-Robot Interaction Using Two-Layer Fuzzy Support Vector Regression [J].
Chen, Lue-Feng ;
Liu, Zhen-Tao ;
Wu, Min ;
Ding, Min ;
Dong, Fang-Yan ;
Hirota, Kaoru .
INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2015, 7 (05) :709-729
[9]   Dynamic Emotion Understanding in Human-Robot Interaction Based on Two-Layer Fuzzy SVR-TS Model [J].
Chen, Luefeng ;
Wu, Min ;
Zhou, Mentian ;
Liu, Zhentao ;
She, Jinhua ;
Hirota, Kaoru .
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2020, 50 (02) :490-501
[10]   Three-Layer Weighted Fuzzy Support Vector Regression for Emotional Intention Understanding in Human Robot Interaction [J].
Chen, Luefeng ;
Zhou, Mengtian ;
Wu, Min ;
She, Jinhua ;
Liu, Zhentao ;
Dong, Fangyan ;
Hirota, Kaoru .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2018, 26 (05) :2524-2538