Convolutional Features-Based Broad Learning With LSTM for Multidimensional Facial Emotion Recognition in Human-Robot Interaction

被引:3
作者
Chen, Luefeng [1 ,2 ]
Li, Min [1 ,2 ]
Wu, Min [1 ,2 ]
Pedrycz, Witold [3 ,4 ,5 ]
Hirota, Kaoru [6 ]
机构
[1] China Univ Geosci, Sch Automat, Hubei Key Lab Adv Control & Intelligent Automat C, Wuhan 430074, Peoples R China
[2] China Univ Geosci, Engn Res Ctr Intelligent Technol Geoexplorat, Minist Educ, Wuhan 430074, Peoples R China
[3] Univ Alberta, Dept Elect & Comp Engn, Edmonton, AB T6G 2R3, Canada
[4] Polish Acad Sci, Syst Res Inst, PL-00901 Warsaw, Poland
[5] Istinye Univ, Dept Comp Engn, TR-34396 Sariyer Istanbul, Turkiye
[6] Tokyo Inst Technol, Tokyo 2268502, Japan
来源
IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS | 2024年 / 54卷 / 01期
基金
中国国家自然科学基金;
关键词
emotion recognition; human-robot interaction; long short-term memory (LSTM); EXPRESSION RECOGNITION; NETWORK; REGRESSION; FRAMEWORK; SYSTEM;
D O I
10.1109/TSMC.2023.3301001
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional feature-based broad learning with long short-term memory (CBLSTM) is proposed to recognize multidimensional facial emotions in human-robot interaction. The CBLSTM model consists of convolution and pooling layers, broad learning (BL), and long-and short-term memory network. It aims to obtain the depth, width, and time scale information of facial emotion through three parts of the model, so as to realize multidimensional facial emotion recognition. CBLSTM adopts the structure of BL after processing was done at the convolution and pooling layer to replace the original random mapping method and extract features with more representation ability, which significantly reduces the computational time of the facial emotion recognition network. Moreover, we adopted incremental learning, which can quickly reconstruct the model without a complete retraining process. Experiments on three databases are developed, including CK+, MMI, and SFEW2.0 databases. The experimental results show that the proposed CBLSTM model using multidimensional information produces higher recognition accuracy than that without time scale information. It is 1.30% higher on the CK+ database and 1.06% higher on the MMI database. The computation time is 9.065 s, which is significantly shorter than the time reported for the convolutional neural network (CNN). In addition, the proposed method obtains improvement compared to the state-of-the-art methods. It improves the recognition rate by 3.97%, 1.77%, and 0.17% compared to that of CNN-SIPS, HOG-TOP, and CMACNN in the CK+ database, 5.17%, 5.14%, and 3.56% compared to TLMOS, ALAW, and DAUGN in the MMI database, and 7.08% and 2.98% compared to CNNVA and QCNN in the SFEW2.0 database.
引用
收藏
页码:64 / 75
页数:12
相关论文
共 50 条
  • [21] Speech Emotion Recognition Using an Enhanced Kernel Isomap for Human-Robot Interaction
    Zhang, Shiqing
    Zhao, Xiaoming
    Lei, Bicheng
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2013, 10
  • [22] Speech emotion recognition in real static and dynamic human-robot interaction scenarios
    Grageda, Nicolas
    Busso, Carlos
    Alvarado, Eduardo
    Garcia, Ricardo
    Mahu, Rodrigo
    Huenupan, Fernando
    Yoma, Nestor Becerra
    COMPUTER SPEECH AND LANGUAGE, 2025, 89
  • [23] Interactive Emotion Recognition Using Support Vector Machine for Human-Robot Interaction
    Tsai, Ching-Chih
    Chen, You-Zhu
    Liao, Ching-Wen
    2009 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC 2009), VOLS 1-9, 2009, : 407 - 412
  • [24] A Sociable Human-robot Interaction Scheme Based on Body Emotion Analysis
    Tehao Zhu
    Zeyang Xia
    Jiaqi Dong
    Qunfei Zhao
    International Journal of Control, Automation and Systems, 2019, 17 : 474 - 485
  • [25] CNN-based Broad Learning with Efficient Incremental Reconstruction Model for Facial Emotion Recognition
    Chen, Luefeng
    Li, Min
    Lai, Xuzhi
    Hirota, Kaoru
    Pedrycz, Witold
    IFAC PAPERSONLINE, 2020, 53 (02): : 10236 - 10241
  • [26] Two-layer fuzzy multiple random forest for speech emotion recognition in human-robot interaction
    Chen, Luefeng
    Su, Wanjuan
    Feng, Yu
    Wu, Min
    She, Jinhua
    Hirota, Kaoru
    INFORMATION SCIENCES, 2020, 509 : 150 - 163
  • [27] A Sociable Human-robot Interaction Scheme Based on Body Emotion Analysis
    Zhu, Tehao
    Xia, Zeyang
    Dong, Jiaqi
    Zhao, Qunfei
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2019, 17 (02) : 474 - 485
  • [28] Enhancing Human-Robot Interaction: Development of Multimodal Robotic Assistant for User Emotion Recognition
    Garcia, Sergio
    Gomez-Donoso, Francisco
    Cazorla, Miguel
    APPLIED SCIENCES-BASEL, 2024, 14 (24):
  • [29] THE DEVELOPMENT OF A FACIAL-AFFECT RECOGNITION SYSTEM FOR APPLICATION IN HUMAN-ROBOT INTERACTION SCENARIOS
    Schacter, David
    Wang, Christopher
    Nejat, Goldie
    Benhabib, Beno
    PROCEEDINGS OF THE ASME INTERNATIONAL DESIGN ENGINEERING TECHNICAL CONFERENCES AND COMPUTERS AND INFORMATION IN ENGINEERING CONFERENCE, 2011, VOL 2, PTS A AND B, 2012, : 865 - 873
  • [30] Facial Communicative Signals Valence Recognition in Task-Oriented Human-Robot Interaction
    Lang, Christian
    Wachsmuth, Sven
    Hanheide, Marc
    Wersing, Heiko
    INTERNATIONAL JOURNAL OF SOCIAL ROBOTICS, 2012, 4 (03) : 249 - 262