A novel dual-channel graph convolutional neural network for facial action unit recognition

被引:8
作者
Jia, Xibin [1 ]
Xu, Shaowu [1 ]
Zhou, Yuhan [1 ]
Wang, Luo [1 ]
Li, Weiting [1 ]
机构
[1] Beijing Univ Technol, Fac Informat Technol, Pingleyuan 100, Beijing 100124, Peoples R China
基金
北京市自然科学基金;
关键词
FACS; GCN; Metric learning; AU relation; Dual-channel graph convolutional neural; network;
D O I
10.1016/j.patrec.2023.01.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial Action Unit (AU) recognition is a challenging problem, where the subtle muscle movement brings diverse AU representations. Recently, AU relations are utilized to assist AU recognition and improve the understanding of AUs. Nevertheless, simply adopting the regular Bayesian networks or the relations between AUs and emotions is not enough for modelling complex AU relations. To provide quantitative measurement of AU relations using the knowledge from FACS, we propose an AU relation quantization autoencoder. Moreover, to cope with the diversity of AUs generated from the individual representation differences and other environmental impacts, we propose a dual-channel graph convolutional neural network (DGCN) obtaining both inherent and random AU relations. The first channel is FACS-based relation graph convolution channel (FACS-GCN) embedding prior knowledge of FACS, and it adjusts the network to the inherent AU dependent relations. The second channel is data-learning-based relation graph convolution channel (DLR-GCN) based on metric learning, and it provides robustness for individual differences and environmental changes. Comprehensive experiments have been conducted on three public datasets: CK+, RAF-AU and DISFA. The experimental results demonstrate that our proposed DGCN can extract the hidden relations well, thereby achieving great performance in AU recognition. (c) 2023 Published by Elsevier B.V.
引用
收藏
页码:61 / 68
页数:8
相关论文
共 34 条
[1]   Learning to transfer: transferring latent task structures and its application to person-specific facial action unit detection [J].
Almaev, Timur ;
Martinez, Brais ;
Valstar, Michel .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :3774-3782
[2]  
Bartlett M. S., 2006, Journal of Multimedia, V1, DOI 10.4304/jmm.1.6.22-35
[3]  
Defferrard M, 2016, ADV NEUR IN, V29
[4]   Compound facial expressions of emotion [J].
Du, Shichuan ;
Tao, Yong ;
Martinez, Aleix M. .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2014, 111 (15) :E1454-E1462
[5]  
Ekman P., 1978, APA PsycTests, DOI DOI 10.1037/T27734-000
[6]  
Friesen W.V., 1983, EMFACS-7: emotional facial action coding system, V2, P1
[7]  
Fung A, 2019, Arxiv, DOI arXiv:1911.05946
[8]  
Garcia V., 2017, Few-shot learning with graph neural networks
[9]  
Gori M, 2005, IEEE IJCNN, P729
[10]  
Henaff M, 2015, Arxiv, DOI arXiv:1506.05163