A Novel Facial Expression Recognition (FER) Model Using Multi-scale Attention Network

被引:0
作者
Ghadai, Chakrapani [1 ]
Patra, Dipti [1 ]
Okade, Manish [1 ]
机构
[1] Natl Inst Technol, Rourkela, India
来源
COMPUTER VISION AND IMAGE PROCESSING, CVIP 2023, PT II | 2024年 / 2010卷
关键词
Facial expression recognition (FER); Deep learning; Muti-scale; Attention; receptive field;
D O I
10.1007/978-3-031-58174-8_29
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial Expression Recognition (FER) faces significant challenges, primarily due to significant variations within classes and subtle visual differences between classes, and limited dataset sizes. Real-world factors such as pose, illumination, and partial occlusion further hinder FER performance. To tackle these challenges, multi-scale and attention-based networks have been widely employed. However, previous approaches have primarily focused on increasing depth while neglecting width, resulting in an inadequate representation of granular facial expression features. This study introduces a novel FER model. A multi-scale attention network (MSA-Net) is designed as a more extensive and deeper network that captures features from various receptive fields through a parallel network structure. Each parallel branch in the proposed network utilizes channel complementary multi-scale blocks, e.g., left multi-scale (MS-L) and right multi-scale (MS-R), to broaden the effective receptive field and capture features having diversity. Additionally, attention networks are employed to emphasize important regions and boost the discriminative capability of the multi-scale features. The performance evaluation of the proposed method was carried out on two popular real-world FER databases: AffectNet and RAF-DB. Our MSA-Net has reduced the impact of the pose, partial occlusions and the network's susceptibility to subtle expression-related variations, thereby outperforming other methods in FER.
引用
收藏
页码:336 / 346
页数:11
相关论文
共 27 条
  • [11] Deep reinforcement learning for robust emotional classification in facial expression recognition
    Li, Huadong
    Xu, Hua
    [J]. KNOWLEDGE-BASED SYSTEMS, 2020, 204 (204)
  • [12] Reliable Crowdsourcing and Deep Locality-Preserving Learning for Expression Recognition in the Wild
    Li, Shan
    Deng, Weihong
    Du, JunPing
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2584 - 2593
  • [13] Learning Informative and Discriminative Features for Facial Expression Recognition in the Wild
    Li, Yingjian
    Lu, Yao
    Chen, Bingzhi
    Zhang, Zheng
    Li, Jinxing
    Lu, Guangming
    Zhang, David
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (05) : 3178 - 3189
  • [14] Occlusion Aware Facial Expression Recognition Using CNN With Attention Mechanism
    Li, Yong
    Zeng, Jiabei
    Shan, Shiguang
    Chen, Xilin
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (05) : 2439 - 2450
  • [15] Facial Expression Recognition with CNN Ensemble
    Liu, Kuang
    Zhang, Minming
    Pan, Zhigeng
    [J]. 2016 INTERNATIONAL CONFERENCE ON CYBERWORLDS (CW), 2016, : 163 - 166
  • [16] AffectNet: A Database for Facial Expression, Valence, and Arousal Computing in the Wild
    Mollahosseini, Ali
    Hasani, Behzad
    Mahoor, Mohammad H.
    [J]. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2019, 10 (01) : 18 - 31
  • [17] Integrating Graph Convolution Into a Deep Multilayer Framework for Low-Light Image Enhancement
    Panda, Santosh Kumar
    Sa, Pankaj Kumar
    [J]. IEEE SENSORS LETTERS, 2024, 8 (05) : 1 - 4
  • [18] Shin M, 2016, IEEE ROMAN, P724, DOI 10.1109/ROMAN.2016.7745199
  • [19] Siqueira H, 2020, AAAI CONF ARTIF INTE, V34, P5800
  • [20] Joint Sparse Learning for 3-D Facial Expression Generation
    Song, Mingli
    Tao, Dacheng
    Sun, Shengpeng
    Chen, Chun
    Bu, Jiajun
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2013, 22 (08) : 3283 - 3295