Method of Multi-Label Visual Emotion Recognition Fusing Fore-Background Features

被引:0
作者
Feng, Yuehua [1 ]
Wei, Ruoyan [1 ]
机构
[1] Hebei Univ Econ & Business, Sch Management Sci & Informat Engn, Shijiazhuang 050061, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 18期
基金
中国国家自然科学基金;
关键词
multi-label; emotion recognition; fore-background features; graph convolutional networks; correlation;
D O I
10.3390/app14188564
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
This paper proposes a method for multi-label visual emotion recognition that fuses fore-background features to address the following issues that visual-based multi-label emotion recognition often overlooks: the impacts of the background that the person is placed in and the foreground, such as social interactions between different individuals on emotion recognition; the simplification of multi-label recognition tasks into multiple binary classification tasks; and it ignores the global correlations between different emotion labels. First, a fore-background-aware emotion recognition model (FB-ER) is proposed, which is a three-branch multi-feature hybrid fusion network. It efficiently extracts body features by designing a core region unit (CR-Unit) that represents background features as background keywords and extracts depth map information to model social interactions between different individuals as foreground features. These three features are fused at both the feature and decision levels. Second, a multi-label emotion recognition classifier (ML-ERC) is proposed, which captures the relationship between different emotion labels by designing a label co-occurrence probability matrix and cosine similarity matrix, and uses graph convolutional networks to learn correlations between different emotion labels to generate a classifier that considers emotion correlations. Finally, the visual features are combined with the object classifier to enable the multi-label recognition of 26 different emotions. The proposed method was evaluated on the Emotic dataset, and the results show an improvement of 0.732% in the mAP and 0.007 in the Jaccard's coefficient compared with the state-of-the-art method.
引用
收藏
页数:21
相关论文
共 40 条
  • [1] [Anonymous], 2019, PROC CVPR IEEE, DOI [DOI 10.48550/ARXIV.1904.03582, DOI 10.1109/CVPR.2019.00532]
  • [2] Emotion Recognition beyond Pixels: Leveraging Facial Point Landmark Meshes
    Arabian, Herag
    Alshirbaji, Tamer Abdulbaki
    Chase, J. Geoffrey
    Moeller, Knut
    [J]. APPLIED SCIENCES-BASEL, 2024, 14 (08):
  • [3] A Mutation in EGF Repeat-8 of Notch Discriminates Between Serrate/Jagged and Delta Family Ligands
    Aviezer, Hillel
    Trope, Yaacov
    Todorov, Alexander
    [J]. SCIENCE, 2012, 338 (6111) : 1225 - 1229
  • [4] Bojanowski P., 2017, Transactions of the ACL, V5, P135, DOI DOI 10.1162/TACL_A_00051
  • [5] Peters ME, 2018, Arxiv, DOI arXiv:1802.05365
  • [6] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [7] Ilyes B., 2020, Inf. Fusion, V76, P422
  • [8] Jinfeng Liu, 2020, Journal of Physics: Conference Series, V1648, DOI 10.1088/1742-6596/1648/2/022112
  • [9] A test of the micro-expressions training tool: Does it improve lie detection?
    Jordan, Sarah
    Brimbal, Laure
    Wallace, D. Brian
    Kassin, Saul M.
    Hartwig, Maria
    Street, Chris N. H.
    [J]. JOURNAL OF INVESTIGATIVE PSYCHOLOGY AND OFFENDER PROFILING, 2019, 16 (03) : 222 - 235
  • [10] FER-BHARAT: a lightweight deep learning network for efficient unimodal facial emotion recognition in Indian context
    Karani R.
    Jani J.
    Desai S.
    [J]. Discover Artificial Intelligence, 4 (1):