Bi-modal emotion recognition from expressive face and body gestures

被引:208
|
作者
Gunes, Hatice
Piccardi, Massimo
机构
[1] Computer Vision Research Group, Faculty of Information Technology, University of Technology, Sydney (UTS), Broadway, NSW, 2007
关键词
Bi-modal emotion recognition; facial expression; expressive body gestures; feature-level fusion; decision-level fusion;
D O I
10.1016/j.jnca.2006.09.007
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Psychological research findings suggest that humans rely on the combined visual channels of face and body more than any other channel when they make judgments about human communicative behavior. However, most of the existing systems attempting to analyze the human nonverbal behavior are mono-modal and focus only on the face. Research that aims to integrate gestures as an expression mean has only recently emerged. Accordingly, this paper presents an approach to automatic visual recognition of expressive face and upper-body gestures from video sequences suitable for use in a vision-based affective multi-modal framework. Face and body movements are captured simultaneously using two separate cameras. For each video sequence single expressive frames both from face and body are selected manually for analysis and recognition of emotions. Firstly, individual classifiers are trained from individual modalities. Secondly, we fuse facial expression and affective body gesture information at the feature and at the decision level. In the experiments performed, the emotion classification using the two modalities achieved a better recognition accuracy outperforming classification using the individual facial or bodily modality alone. (c) 2006 Elsevier Ltd. All rights reserved.
引用
收藏
页码:1334 / 1345
页数:12
相关论文
共 50 条
  • [31] Bi-modal annoyance level detection from speech and text
    Justo, Raquel
    Irastorza, Jon
    Perez, Saioa
    Ines Torres, M.
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2018, (61): : 83 - 89
  • [32] Multimodal Emotion Recognition Based on Facial Expressions, Speech, and Body Gestures
    Yan, Jingjie
    Li, Peiyuan
    Du, Chengkun
    Zhu, Kang
    Zhou, Xiaoyang
    Liu, Ying
    Wei, Jinsheng
    ELECTRONICS, 2024, 13 (18)
  • [33] A sketch recognition method based on bi-modal model using cooperative learning paradigm
    Zhang S.
    Wang L.
    Cui Z.
    Wang S.
    Neural Computing and Applications, 2024, 36 (23) : 14275 - 14290
  • [34] State synchronous modeling of audio-visual information for bi-modal speech recognition
    Nakamura, S
    Kumatani, K
    Tamura, S
    ASRU 2001: IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, CONFERENCE PROCEEDINGS, 2001, : 409 - 412
  • [35] STRUCTURAL ANALYSIS OF BODY UNDER 3-POINTS SUPPORTING FOR BI-MODAL TRAM
    Kim, Yeon-Su
    Lee, Kang-Won
    Mok, Jai-Kyun
    Leem, Song-Gyu
    Jeong, Jong-Cheol
    ADVANCES IN FRACTURE AND DAMAGE MECHANICS VIII, 2010, 417-418 : 401 - +
  • [36] Emotion Recognition in Face and Body Motion in Bulimia Nervosa
    Dapelo, Marcela Marin
    Surguladze, Simon
    Morris, Robin
    Tchanturia, Kate
    EUROPEAN EATING DISORDERS REVIEW, 2017, 25 (06) : 595 - 600
  • [37] THE FUSION KNOWLEDGE OF FACE, BODY AND CONTEXT FOR EMOTION RECOGNITION
    Wu, Jingjing
    Zhang, Yong
    Ning, Li
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA & EXPO WORKSHOPS (ICMEW), 2019, : 108 - 113
  • [38] From Face Identification to Emotion Recognition
    Chang, Xin
    Skarbek, Wladyslaw
    PHOTONICS APPLICATIONS IN ASTRONOMY, COMMUNICATIONS, INDUSTRY, AND HIGH-ENERGY PHYSICS EXPERIMENTS 2019, 2019, 11176
  • [39] Decoding Emotion in Drug Abusers: Evidence for Face and Body Emotion Recognition and for Disgust Emotion
    Bonfiglio, Natale Salvatore
    Renati, Roberta
    Bottini, Gabriella
    EUROPEAN JOURNAL OF INVESTIGATION IN HEALTH PSYCHOLOGY AND EDUCATION, 2022, 12 (09) : 1427 - 1440
  • [40] Robust bi-modal speech recognition based on state synchronous modeling and stream weight optimization
    Nakamura, S
    Kumatani, K
    Tamura, S
    2002 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOLS I-IV, PROCEEDINGS, 2002, : 309 - 312