Multi-modal facial expression feature based on deep-neural networks

被引:0
作者
Wei Wei
Qingxuan Jia
Yongli Feng
Gang Chen
Ming Chu
机构
[1] Beijing Institute of Petrochemical Technology,College of Information Engineering
[2] Beijing University of Posts and Telecommunications,School of Automation
来源
Journal on Multimodal User Interfaces | 2020年 / 14卷
关键词
Emotion recognition; Multi-modal feature; Convolutional neural networks; Support vector machine;
D O I
暂无
中图分类号
学科分类号
摘要
Emotion recognition based on facial expression is a challenging research topic and has attracted a great deal of attention in the past few years. This paper presents a novel method, utilizing multi-modal strategy to extract emotion features from facial expression images. The basic idea is to combine the low-level empirical feature and the high-level self-learning feature into a multi-modal feature. The 2-dimensional coordinate of facial key points are extracted as low-level empirical feature and the high-level self-learning feature are extracted by the Convolutional Neural Networks (CNNs). To reduce the number of free parameters of CNNs, small filters are utilized for all convolutional layers. Owing to multiple small filters are equivalent of a large filter, which can reduce the number of parameters to learn effectively. And label-preserving transformation is used to enlarge the dataset artificially, in order to address the over-fitting and data imbalance of deep neural networks. Then, two kinds of modal features are fused linearly to form the facial expression feature. Extensive experiments are evaluated on the extended Cohn–Kanade (CK+) Dataset. For comparison, three kinds of feature vectors are adopted: low-level facial key point feature vector, high-level self-learning feature vector and multi-modal feature vector. The experiment results show that the multi-modal strategy can achieve encouraging recognition results compared to the single modal strategy.
引用
收藏
页码:17 / 23
页数:6
相关论文
共 87 条
  • [21] Huang Z(2001)Random forests Mach Learn 45 5-32
  • [22] Gao XB(1998)A model of saliency-based visual attention for rapid sceneanalysis IEEE Trans Pattern Anal Mach Intell 20 1254-1259
  • [23] Su Y(1997)Multitask learning Chine Learn 28 41-75
  • [24] Li XL(2015)Effective facial expression recognition via the boosted convolutional neural network Commun Comput Inf Sci 546 179-188
  • [25] Tao DC(2003)Automatic facial expression analysis: a survey Pattern Recognit 36 259-275
  • [26] Sung JW(2015)Automatic facial expression recognition using features of salient facial patches IEEE Trans Affect Comput 6 1-12
  • [27] Kanada T(2015)Automatic facial expression recognition using features of salient facial patches IEEE Trans Affect Comput 6 1-12
  • [28] Kim DJ(2015)AU-inspired deep networks for facial expression feature learning Neurocomputing 159 126-136
  • [29] Chen SZ(undefined)undefined undefined undefined undefined-undefined
  • [30] Wang HP(undefined)undefined undefined undefined undefined-undefined