Harmonious Mutual Learning for Facial Emotion Recognition

被引:0
作者
Yanling Gan
Luhui Xu
Haiying Xia
Gan Liu
机构
[1] Guangxi Normal University,School of Computer Science and Engineering
[2] Guangxi Normal University,School of Electronic and Information Engineering
来源
Neural Processing Letters | / 56卷
关键词
Facial emotion recognition; Mutual learning; Attention mechanisms; Probability distributions;
D O I
暂无
中图分类号
学科分类号
摘要
Facial emotion recognition in the wild is an important task in computer vision, but it still remains challenging since the influence of backgrounds, occlusions and illumination variations in facial images, as well as the ambiguity of expressions. This paper proposes a harmonious mutual learning framework for emotion recognition, mainly through utilizing attention mechanisms and probability distributions without utilizing additional information. Specifically, this paper builds an architecture with two emotion recognition networks and makes progressive cooperation and interaction between them. We first integrate self-mutual attention module into the backbone to learn discriminative features against the influence from emotion-irrelevant facial information. In this process, we deploy spatial attention module and convolutional block attention module for the two networks respectively, guiding to enhanced and supplementary learning of attention. Further, in the classification head, we propose to learn the latent ground-truth emotion probability distributions using softmax function with temperature to characterize the expression ambiguity. On this basis, a probability distribution distillation learning module is constructed to perform class semantic interaction using bi-directional KL loss, allowing mutual calibration for the two networks. Experimental results on three public datasets show the superiority of the proposed method compared to state-of-the-art ones.
引用
收藏
相关论文
共 106 条
  • [1] Sajjad M(2020)Human behavior understanding in big multimedia data using cnn based facial expression recognition Mobile Netw Appl 25 1611-1621
  • [2] Zahir S(2022)Classifying emotions and engagement in online learning based on a single facial expression recognition neural network IEEE Trans Affect Comput 13 2132-2143
  • [3] Ullah A(2020)Multiple attention network for facial expression recognition IEEE Access 8 7383-7393
  • [4] Akhtar Z(2023)Distract your attention: multi-head cross attention network for facial expression recognition Biomimetics 8 199-9562
  • [5] Muhammad K(2023)Joint spatial and scale attention network for multi-view facial expression recognition Pattern Recognit. 139 9543-121
  • [6] Savchenko AV(2023)A framework for facial expression recognition using deep self-attention network J Ambient Intell Human Comput 14 113-2139
  • [7] Savchenko LV(2022)Handwritten mathematical expression recognition via attention aggregation based bi-directional mutual learning Proc the AAAI Conf Artif Intell 36 2127-5373
  • [8] Makarov I(2019)Global-local mutual attention model for text classification IEEE/ACM Trans Audio Speech Lang Process 27 5361-4203
  • [9] Gan Y(2022)Dual mutual learning for cross-modality person re-identification IEEE Trans Circuits Syst Video Technol 32 4193-12
  • [10] Chen J(2023)Learn from each other to classify better: cross-layer mutual attention learning for fine-grained visual classification Pattern Recognit 140 1-213