Self Decoupling-Reconstruction Network for Facial Expression Recognition

被引:0
作者
Wang, Linhuang [1 ]
Kang, Xin [1 ]
Ding, Fei [1 ]
Yu, Hai-tao [2 ]
Wu, Yunong [3 ]
Matsumoto, Kazuyuki [1 ]
Nakagawa, Satoshi [4 ]
Ren, Fuji [5 ]
机构
[1] Tokushima Univ, Grad Sch Adv Technol & Sci, Tokushima, Japan
[2] Univ Tsukuba, Fac Lib Informat & Media Sci, Tsukuba, Ibaraki, Japan
[3] Dataa Robot, Chengdu, Peoples R China
[4] Univ Tokyo, Grad Sch Informat Sci & Technol, Tokyo, Japan
[5] Univ Elect Sci & Technol China, Sch Comp Sci & Engn, Chengdu, Peoples R China
来源
2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024 | 2024年
关键词
Facial Expression Recognition; Image Reconstruction; Affective Computing; Convolution Neural Network;
D O I
10.1109/IJCNN60899.2024.10651392
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial Expression Recognition (FER) poses significant challenges due to various imaging conditions, including diverse head poses, lighting conditions, resolutions, and occlusions. Additionally, different personal attributes such as age, gender, and racial background further contribute to the complexity of FER. To accurately extract meaningful expression features amidst these interfering factors to enhance recognition accuracy and the model's generalization, we propose a Self DecouplingReconstruction Network (SDRNet). Specifically, our approach involves two learning processes. In the first phase, the network is trained to decouple facial images with expressions into expression and neutral components. This process involves reconstructing neutral facial images and the original input, ensuring the preservation of meaningful expression components devoid of interference in the decoupling process. In the second learning phase, we employ simple convolutional neural networks (CNNs) to recognize the extracted expression components. Our method has achieved state-of-the-art results across multiple widely used datasets, providing substantial evidence of its effectiveness. Additionally, we demonstrate the robust generalization performance of our approach through cross-database evaluations.
引用
收藏
页数:8
相关论文
共 41 条
[1]   RADIOGAN:Deep Convolutional Conditional Generative Adversarial Network to Generate PET Images [J].
Amyar, Amine ;
Ruan, Su ;
Verra, Pierre ;
Decazes, Pierre ;
Modzelewski, Romain .
2020 7TH INTERNATIONAL CONFERENCE ON BIOINFORMATICS RESEARCH AND APPLICATIONS, ICBRA 2020, 2020, :28-33
[2]  
Calder Andrew J, 2016, Facial Expression Recognition, P41
[3]   Dual subspace manifold learning based on GCN for intensity-invariant facial expression recognition [J].
Chen, Jingying ;
Shi, Jinxin ;
Xu, Ruyi .
PATTERN RECOGNITION, 2024, 148
[4]  
Darwin C., 1872, P374
[5]   RetinaFace: Single-shot Multi-level Face Localisation in the Wild [J].
Deng, Jiankang ;
Guo, Jia ;
Ververas, Evangelos ;
Kotsia, Irene ;
Zafeiriou, Stefanos .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, :5202-5211
[6]   FaceNet2ExpNet: Regularizing a Deep Face Recognition Net for Expression Recognition [J].
Ding, Hui ;
Zhou, Shaohua Kevin ;
Chellappa, Rama .
2017 12TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2017), 2017, :118-126
[7]  
Dosovitskiy A., 2020, arXiv, V2010, P11929, DOI [10.48550/arXiv.2010.11929Focustolearnmore, DOI 10.48550/ARXIV.2010.11929FOCUSTOLEARNMORE]
[8]  
Guo YM, 2012, LECT NOTES COMPUT SC, V7573, P631, DOI 10.1007/978-3-642-33709-3_45
[9]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[10]  
Hu Y, 2008, Pipelines 2008: Pipeline Asset Management: Maximizing Performance of Our Pipeline Infrastructure, P1, DOI 10.1109/AFGR.2008.4813445