From Facial Expression Recognition to Interpersonal Relation Prediction

被引:168
作者
Zhang, Zhanpeng [1 ]
Luo, Ping [2 ]
Loy, Chen Change [2 ]
Tang, Xiaoou [2 ]
机构
[1] SenseTime Grp Ltd, Shatin, Hong Kong, Peoples R China
[2] Chinese Univ Hong Kong, Dept Informat Engn, Shatin, Hong Kong, Peoples R China
关键词
Facial expression recognition; Interpersonal relation; Deep convolutional network; IMAGES;
D O I
10.1007/s11263-017-1055-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Interpersonal relation defines the association, e.g., warm, friendliness, and dominance, between two or more people. We investigate if such fine-grained and high-level relation traits can be characterized and quantified from face images in the wild. We address this challenging problem by first studying a deep network architecture for robust recognition of facial expressions. Unlike existing models that typically learn from facial expression labels alone, we devise an effective multitask network that is capable of learning from rich auxiliary attributes such as gender, age, and head pose, beyond just facial expression data. While conventional supervised training requires datasets with complete labels (e.g., all samples must be labeled with gender, age, and expression), we show that this requirement can be relaxed via a novel attribute propagation method. The approach further allows us to leverage the inherent correspondences between heterogeneous attribute sources despite the disparate distributions of different datasets. With the network we demonstrate state-of-the-art results on existing facial expression recognition benchmarks. To predict inter-personal relation, we use the expression recognition network as branches for a Siamese model. Extensive experiments show that our model is capable of mining mutual context of faces for accurate fine-grained interpersonal prediction.
引用
收藏
页码:550 / 569
页数:20
相关论文
共 97 条
  • [51] Goodfellow I., 2013, Challenges in representation learning: A report on three machine learning contests
  • [52] Gottman J., 2001, J FAM COMMUN, V7, P37, DOI DOI 10.1207/S15327698JFC0101_06
  • [53] The influence of facial emotion displays, gender, and ethnicity on judgments of dominance and affiliation
    Hess, U
    Blairy, S
    Kleck, RE
    [J]. JOURNAL OF NONVERBAL BEHAVIOR, 2000, 24 (04) : 265 - 283
  • [54] Deep Learning for Emotion Recognition on Small Datasets Using Transfer Learning
    Hong-Wei Ng
    Viet Dung Nguyen
    Vonikakis, Vassilios
    Winkler, Stefan
    [J]. ICMI'15: PROCEEDINGS OF THE 2015 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2015, : 443 - 449
  • [55] EPPD: Efficient and Privacy-Preserving Proximity Testing with Differential Privacy Techniques
    Huang, Cheng
    Lu, Rongxing
    Zhu, Hui
    Shao, Jun
    Alamer, Abdulrahman
    Lin, Xiaodong
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2016,
  • [56] Huang S, 2016, IEEE IPCCC
  • [57] Hung H., 2007, ACM Multimedia
  • [58] Visual Persuasion: Inferring Communicative Intents of Images
    Joo, Jungseock
    Li, Weixin
    Steen, Francis F.
    Zhu, Song-Chun
    [J]. 2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, : 216 - 223
  • [59] Joint Fine-Tuning in Deep Neural Networks for Facial Expression Recognition
    Jung, Heechul
    Lee, Sihaeng
    Yim, Junho
    Park, Sunjeong
    Kim, Junmo
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 2983 - 2991
  • [60] THE 1982 INTERPERSONAL CIRCLE - A TAXONOMY FOR COMPLEMENTARITY IN HUMAN TRANSACTIONS
    KIESLER, DJ
    [J]. PSYCHOLOGICAL REVIEW, 1983, 90 (03) : 185 - 214