Patch-Aware Representation Learning for Facial Expression Recognition

被引:0
作者
Wu, Yi [1 ]
Wang, Shangfei [1 ]
Chang, Yanan [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Anhui, Peoples R China
来源
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023 | 2023年
基金
国家重点研发计划;
关键词
patch-aware; two collaborative streams; facial landmarks; facial expression recognition; JOINT; POSE;
D O I
10.1145/3581783.3612342
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing methods for facial expression recognition (FER) lack the utilization of prior facial knowledge, primarily focusing on expression-related regions while disregarding explicitly processing expression-independent information. This paper proposes a patch-aware FER method that incorporates facial keypoints to guide the model and learns precise representations through two collaborative streams, addressing these issues. First, facial keypoints are detected using a facial landmark detection algorithm, and the facial image is divided into equal-sized patches using the Patch Embedding Module. Then, a correlation is established between the keypoints and patches using a simplified conversion relationship. Two collaborative streams are introduced, each corresponding to a specific mask strategy. The first stream masks patches corresponding to the keypoints, excluding those along the facial contour, with a certain probability. The resulting image embedding is input into the Encoder to obtain expression-related features. The features are passed through the Decoder and Classifier to reconstruct the masked patches and recognize the expression, respectively. The second stream masks patches corresponding to all the above keypoints. The resulting image embedding is input into the Encoder and Classifier successively, with the resulting logit approximating a uniform distribution. Through the first stream, the Encoder learns features in the regions related to expression, while the second stream enables the Encoder to better ignore expression-independent information, such as the background, facial contours, and hair. Experiments on two bench-mark datasets demonstrate that the proposed method outperforms state-of-the-art methods.
引用
收藏
页码:6143 / 6151
页数:9
相关论文
共 50 条
  • [21] Discriminative feature learning-based pixel difference representation for facial expression recognition
    Sun, Zhe
    Hu, Zheng-Ping
    Wang, Meng
    Zhao, Shu-Huan
    IET COMPUTER VISION, 2017, 11 (08) : 675 - 682
  • [22] Multimodal learning for facial expression recognition
    Zhang, Wei
    Zhang, Youmei
    Ma, Lin
    Guan, Jingwei
    Gong, Shijie
    PATTERN RECOGNITION, 2015, 48 (10) : 3191 - 3202
  • [23] Joint Patch and Multi-label Learning for Facial Action Unit and Holistic Expression Recognition
    Zhao, Kaili
    Chu, Wen-Sheng
    De la Torre, Fernando
    Cohn, Jeffrey F.
    Zhang, Honggang
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (08) : 3931 - 3946
  • [24] A Comparison of Facial Feature Representation Methods for Automatic Facial Expression Recognition
    Deaney, Waleed
    Venter, Isabella
    Ghaziasgar, Mehrdad
    Dodds, Reg
    SOUTH AFRICAN INSTITUTE OF COMPUTER SCIENTISTS AND INFORMATION TECHNOLOGISTS (SACSIT 2017), 2017, : 85 - 94
  • [25] Facial Expression Recognition Based on Fusion of Sparse Representation
    Ying, Zi-Lu
    Wang, Zhe-Wei
    Huang, Ming-Wei
    ADVANCED INTELLIGENT COMPUTING THEORIES AND APPLICATIONS: WITH ASPECTS OF ARTIFICIAL INTELLIGENCE, 2010, 6216 : 457 - 464
  • [26] Pose-Aware Facial Expression Recognition Assisted by Expression Descriptions
    Wang, Shangfei
    Wu, Yi
    Chang, Yanan
    Li, Guoming
    Mao, Meng
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (01) : 241 - 253
  • [27] A joint learning method with consistency-aware for low-resolution facial expression recognition
    Xie, Yuanlun
    Tian, Wenhong
    Song, Liang
    Xue, Ruini
    Zha, Zhiyuan
    Wen, Bihan
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 244
  • [28] Learning facial expression-aware global-to-local representation for robust action unit detection
    Rudong An
    Aobo Jin
    Wei Chen
    Wei Zhang
    Hao Zeng
    Zhigang Deng
    Yu Ding
    Applied Intelligence, 2024, 54 : 1405 - 1425
  • [29] Learning facial expression-aware global-to-local representation for robust action unit detection
    An, Rudong
    Jin, Aobo
    Chen, Wei
    Zhang, Wei
    Zeng, Hao
    Deng, Zhigang
    Ding, Yu
    APPLIED INTELLIGENCE, 2024, 54 (02) : 1405 - 1425
  • [30] Deep Margin-Sensitive Representation Learning for Cross-Domain Facial Expression Recognition
    Li, Yingjian
    Zhang, Zheng
    Chen, Bingzhi
    Lu, Guangming
    Zhang, David
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 1359 - 1373