Decision Level Fusion of Domain Specific Regions for Facial Action Recognition

被引:33
作者
Jiang, Bihan [1 ]
Martinez, Brais [1 ]
Valstar, Michel F. [2 ]
Pantic, Maja [1 ,3 ]
机构
[1] Univ London Imperial Coll Sci Technol & Med, Dept Comp, London SW7 2AZ, England
[2] Univ Nottingham, Sch Comp Sci, Mixed Real Lab, Nottingham NG7 2RD, England
[3] Univ Twente, Fac Elect Engn Math & Comp Sci, NL-7500 AE Enschede, Netherlands
来源
2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR) | 2014年
基金
英国工程与自然科学研究理事会;
关键词
EXPRESSION RECOGNITION; FEATURES;
D O I
10.1109/ICPR.2014.312
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper we propose a new method for the detection of action units that relies on a novel region-based face representation and a mid-level decision layer that combines region-specific information. Different from other approaches, we do not represent the face as a regular grid based on the face location alone (holistic representation), nor by using small patches centred at fiducial facial point locations (local representation). Instead, we propose to use domain knowledge regarding AU-specific facial muscle contractions to define a set of face regions covering the whole face. Therefore, as opposed to local appearance models, our face representation makes use of the full facial appearance, while the use of facial point locations to define the regions means that we obtain better-registered descriptors compared to holistic representations. Finally, we propose an AU-specific weighted sum model is used as a decision-level fusion layer in charge of combining region-specific probabilistic information. This configuration allows each classifier to learning the typical appearance changes for a specific face part and reduces the dimensionality of the problem thus proving to be more robust. Our approach is evaluated on the DISFA and GEMEP-FERA datasets using two histogram-based appearance features, Local Binary Pattern and Local Phase Quantisation. We show superior performance for both the domain-specific region definition and the decision-level fusion respect to the standard approaches when it comes to automatic facial action unit detection.
引用
收藏
页码:1776 / 1781
页数:6
相关论文
共 24 条
[1]  
Almaev T., 2013, INT C AFF COMP INT I
[2]   All Smiles are Not Created Equal: Morphology and Timing of Smiles Perceived as Amused, Polite, and Embarrassed/Nervous [J].
Ambadar, Zara ;
Cohn, Jeffrey F. ;
Reed, Lawrence Ian .
JOURNAL OF NONVERBAL BEHAVIOR, 2009, 33 (01) :17-34
[3]  
[Anonymous], 2002, Manual and Investigators Guide
[4]  
[Anonymous], 2011, IEEE INT C AUT FAC G
[5]  
[Anonymous], KAHIERS PSYCHIATRIQU
[6]  
[Anonymous], 1988, Signal Detection Theory and Psychophysics
[7]  
[Anonymous], 2012, AS C COMP VIS
[8]  
Bartlett M. S., 2006, Journal of Multimedia, V1, DOI 10.4304/jmm.1.6.22-35
[9]  
Bihan Jiang, 2011, Proceedings 2011 IEEE International Conference on Automatic Face & Gesture Recognition (FG 2011), P314, DOI 10.1109/FG.2011.5771416
[10]   In the Pursuit of Effective Affective Computing: The Relationship Between Features and Registration [J].
Chew, S. W. ;
Lucey, P. ;
Lucey, S. ;
Saragih, J. ;
Cohn, J. F. ;
Matthews, I. ;
Sridharan, S. .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2012, 42 (04) :1006-1016