Multimodal emotion recognition based on peak frame selection from video

被引:24
|
作者
Zhalehpour, Sara [1 ]
Akhtar, Zahid [2 ]
Erdem, Cigdem Eroglu [3 ]
机构
[1] INRS EMT, Montreal, PQ, Canada
[2] Univ Udine, I-33100 Udine, Italy
[3] Bahcesehir Univ, Istanbul, Turkey
关键词
Affective computing; Facial expression recognition; Apex frame; Audio-visual emotion recognition; FUSION;
D O I
10.1007/s11760-015-0822-0
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We present a fully automatic multimodal emotion recognition system based on three novel peak frame selection approaches using the video channel. Selection of peak frames (i.e., apex frames) is an important preprocessing step for facial expression recognition as they contain the most relevant information for classification. Two of the three proposed peak frame selection methods (i.e., MAXDIST and DEND-CLUSTER) do not employ any training or prior learning. The third method proposed for peak frame selection (i.e., EIFS) is based on measuring the "distance" of the expressive face from the subspace of neutral facial expression, which requires a prior learning step to model the subspace of neutral face shapes. The audio and video modalities are fused at the decision level. The subject-independent audio-visual emotion recognition system has shown promising results on two databases in two different languages (eNTERFACE and BAUM-1a).
引用
收藏
页码:827 / 834
页数:8
相关论文
共 50 条
  • [21] Efficient video face recognition based on frame selection and quality assessment
    Kharchevnikova A.
    Savchenko A.V.
    PeerJ Computer Science, 2021, 7 : 1 - 18
  • [22] Emotion Recognition Based on Multimodal Information
    Zeng, Zhihong
    Pantic, Maja
    Huang, Thomas S.
    AFFECTIVE INFORMATION PROCESSING, 2009, : 241 - +
  • [23] TEMPORALLY CONSISTENT KEY FRAME SELECTION FROM VIDEO FOR FACE RECOGNITION
    Saeed, Usman
    Dugelay, Jean-Luc
    18TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO-2010), 2010, : 1311 - 1315
  • [24] Video Multimodal Emotion Recognition System for Real World Applications
    Lee, Sun-Kyung
    Kim, Jong-Hwan
    INTERSPEECH 2023, 2023, : 668 - 669
  • [25] Multimodal Dimensional and Continuous Emotion Recognition in Dyadic Video Interactions
    Zhao, Jinming
    Chen, Shizhe
    Jin, Qin
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING, PT I, 2018, 11164 : 301 - 312
  • [26] Audio-Video Based Multimodal Emotion Recognition Using SVMs and Deep Learning
    Sun, Bo
    Xu, Qihua
    He, Jun
    Yu, Lejun
    Li, Liandong
    Wei, Qinglan
    PATTERN RECOGNITION (CCPR 2016), PT II, 2016, 663 : 621 - 631
  • [27] Multimodal interaction enhanced representation learning for video emotion recognition
    Xia, Xiaohan
    Zhao, Yong
    Jiang, Dongmei
    FRONTIERS IN NEUROSCIENCE, 2022, 16
  • [28] A multimodal emotion recognition model integrating speech, video and MoCAP
    Ning Jia
    Chunjun Zheng
    Wei Sun
    Multimedia Tools and Applications, 2022, 81 : 32265 - 32286
  • [29] EmoNets: Multimodal deep learning approaches for emotion recognition in video
    Kahou, Samira Ebrahimi
    Bouthillier, Xavier
    Lamblin, Pascal
    Gulcehre, Caglar
    Michalski, Vincent
    Konda, Kishore
    Jean, Sebastien
    Froumenty, Pierre
    Dauphin, Yann
    Boulanger-Lewandowski, Nicolas
    Ferrari, Raul Chandias
    Mirza, Mehdi
    Warde-Farley, David
    Courville, Aaron
    Vincent, Pascal
    Memisevic, Roland
    Pal, Christopher
    Bengio, Yoshua
    JOURNAL ON MULTIMODAL USER INTERFACES, 2016, 10 (02) : 99 - 111
  • [30] EmoNets: Multimodal deep learning approaches for emotion recognition in video
    Samira Ebrahimi Kahou
    Xavier Bouthillier
    Pascal Lamblin
    Caglar Gulcehre
    Vincent Michalski
    Kishore Konda
    Sébastien Jean
    Pierre Froumenty
    Yann Dauphin
    Nicolas Boulanger-Lewandowski
    Raul Chandias Ferrari
    Mehdi Mirza
    David Warde-Farley
    Aaron Courville
    Pascal Vincent
    Roland Memisevic
    Christopher Pal
    Yoshua Bengio
    Journal on Multimodal User Interfaces, 2016, 10 : 99 - 111