Multi-modal facial expression feature based on deep-neural networks

被引:0
作者
Wei Wei
Qingxuan Jia
Yongli Feng
Gang Chen
Ming Chu
机构
[1] Beijing Institute of Petrochemical Technology,College of Information Engineering
[2] Beijing University of Posts and Telecommunications,School of Automation
来源
Journal on Multimodal User Interfaces | 2020年 / 14卷
关键词
Emotion recognition; Multi-modal feature; Convolutional neural networks; Support vector machine;
D O I
暂无
中图分类号
学科分类号
摘要
Emotion recognition based on facial expression is a challenging research topic and has attracted a great deal of attention in the past few years. This paper presents a novel method, utilizing multi-modal strategy to extract emotion features from facial expression images. The basic idea is to combine the low-level empirical feature and the high-level self-learning feature into a multi-modal feature. The 2-dimensional coordinate of facial key points are extracted as low-level empirical feature and the high-level self-learning feature are extracted by the Convolutional Neural Networks (CNNs). To reduce the number of free parameters of CNNs, small filters are utilized for all convolutional layers. Owing to multiple small filters are equivalent of a large filter, which can reduce the number of parameters to learn effectively. And label-preserving transformation is used to enlarge the dataset artificially, in order to address the over-fitting and data imbalance of deep neural networks. Then, two kinds of modal features are fused linearly to form the facial expression feature. Extensive experiments are evaluated on the extended Cohn–Kanade (CK+) Dataset. For comparison, three kinds of feature vectors are adopted: low-level facial key point feature vector, high-level self-learning feature vector and multi-modal feature vector. The experiment results show that the multi-modal strategy can achieve encouraging recognition results compared to the single modal strategy.
引用
收藏
页码:17 / 23
页数:6
相关论文
共 87 条
  • [1] Zhao XM(2016)Outcome facial expression recognition: feature extraction and classification IETE Tech Rev 33 505-517
  • [2] Zhang SQ(2015)Robust face recognition via multimodal deep face representation IEEE Trans Multimed 17 2049-2058
  • [3] Ding CX(2015)Neutral face classification using personalized appearance models for fast and robust emotion detection IEEE Trans Image Process 24 2701-2711
  • [4] Tao DC(2016)Pose-robust feature learning for facial expression recognition Front Comput Sci 10 832-844
  • [5] Shi XS(2007)Dynamic texture recognition using local binary patterns with an application to facial expressions IEEE Trans Pattern Anal Mach Intell 29 915-928
  • [6] Guo ZH(1987)An evaluation of the two-dimensional gabor filter model of simple receptive fields in cat striate cortex J Neurophysiol 58 1233-1258
  • [7] Nie FP(2015)Facial expression recognition based on AAM-SIFT and adaptive regional weighting IEEJ Trans Electr Electron Eng 10 713-722
  • [8] Yang L(2010)A review of active appearance models IEEE Trans Syst Man Cybern Part C Appl Rev 40 145-158
  • [9] You J(2007)A unified gradient-based approach for combining ASM into AAM Int J Comput Vis 75 297-310
  • [10] Tao DC(2016)Target classification using the deep convolutional networks for SAR images IEEE Trans Geosci Remote Sens 58 4806-4817