UA-FER: Uncertainty-aware representation learning for facial expression recognition

被引:1
作者
Zhou, Haoliang [1 ]
Huang, Shucheng [1 ]
Xu, Yuqiao [2 ]
机构
[1] Jiangsu Univ Sci & Technol, Sch Comp, Zhenjiang 212003, Peoples R China
[2] Tianjin Univ Technol, Sch Comp Sci & Engn, Tianjin 300384, Peoples R China
基金
中国国家自然科学基金;
关键词
Facial expression recognition; Uncertainty-aware representation learning; Evidential deep learning; Vision-language pre-training model; Knowledge distillation; FEATURES;
D O I
10.1016/j.neucom.2024.129261
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial Expression Recognition (FER) remains a challenging task due to unconstrained conditions like variations in illumination, pose, and occlusion. Current FER approaches mainly focus on learning discriminative features through local attention and global perception of visual encoders, while neglecting the rich semantic information in the text modality. Additionally, these methods rely solely on the softmax-based activation layer for predictions, resulting in overconfident decision-making that hampers the effective handling of uncertain samples and relationships. Such insufficient representations and overconfident predictions degrade recognition performance, particularly in unconstrained scenarios. To tackle these issues, we propose an end-to-end FER framework called UA-FER, which integrates vision-language pre-training (VLP) models with evidential deep learning (EDL) theory to enhance recognition accuracy and robustness. Specifically, to identify multi-grained discriminative regions, we propose the Multi-granularity Feature Decoupling (MFD) module, which decouples global and local facial representations based on image-text affinity while distilling the universal knowledge from the pre-trained VLP models. Additionally, to mitigate misjudgments in uncertain visual-textual relationships, we introduce the Relation Uncertainty Calibration (RUC) module, which corrects these uncertainties using EDL theory. In this way, the model enhances its ability to capture emotion-related discriminative representations and tackle uncertain relationships, thereby improving overall recognition accuracy and robustness. Extensive experiments on in-the-wild and in-the-lab datasets demonstrate that our UA-FER outperforms the state-of-the-art models.
引用
收藏
页数:13
相关论文
共 77 条
  • [31] Adaptively Learning Facial Expression Representation via C-F Labels and Distillation
    Li, Hangyu
    Wang, Nannan
    Ding, Xinpeng
    Yang, Xi
    Gao, Xinbo
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2021, 30 (30) : 2016 - 2028
  • [32] Li HT, 2023, Arxiv, DOI arXiv:2303.00193
  • [33] Reliable Crowdsourcing and Deep Locality-Preserving Learning for Expression Recognition in the Wild
    Li, Shan
    Deng, Weihong
    Du, JunPing
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2584 - 2593
  • [34] Li Y, 2021, arXiv
  • [35] Occlusion Aware Facial Expression Recognition Using CNN With Attention Mechanism
    Li, Yong
    Zeng, Jiabei
    Shan, Shiguang
    Chen, Xilin
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (05) : 2439 - 2450
  • [36] Adaptive Multilayer Perceptual Attention Network for Facial Expression Recognition
    Liu, Hanwei
    Cai, Huiling
    Li, Qingcheng
    Li, Xuefeng
    Xiao, Hui
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (09) : 6253 - 6266
  • [37] Liu LP, 2008, STUD FUZZ SOFT COMP, V219, P1, DOI 10.1007/978-3-540-44792-4
  • [38] Joint spatial and scale attention network for multi-view facial expression recognition
    Liu, Yuanyuan
    Peng, Jiyao
    Dai, Wei
    Zeng, Jiabei
    Shan, Shiguang
    [J]. PATTERN RECOGNITION, 2023, 139
  • [39] Teaching with Soft Label Smoothing for Mitigating Noisy Labels in Facial Expressions
    Lukov, Tohar
    Zhao, Na
    Lee, Gim Hee
    Lim, Ser-Nam
    [J]. COMPUTER VISION, ECCV 2022, PT XII, 2022, 13672 : 648 - 665
  • [40] Facial Expression Recognition With Visual Transformers and Attentional Selective Fusion
    Ma, Fuyan
    Sun, Bin
    Li, Shutao
    [J]. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (02) : 1236 - 1248