Latent-Based Adversarial Neural Networks for Facial Affect Estimations

被引:6
作者
Aspandi, Decky [1 ,2 ]
Mallol-Ragolta, Adria [2 ]
Schuller, Bjoern [2 ,3 ]
Binefa, Xavier [1 ]
机构
[1] Pompeu Fabra Univ, Dept Informat & Commun Technol, Barcelona, Spain
[2] Univ Augsburg, Chair Embedded Intelligence Hlth Care & Wellbeing, Augsburg, Germany
[3] Imperial Coll London, GLAM Grp Language Audio & Mus, London, England
来源
2020 15TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2020) | 2020年
基金
欧盟地平线“2020”;
关键词
D O I
10.1109/FG47880.2020.00053
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There is a growing interest in affective computing research nowadays given its crucial role in bridging humans with computers. This progress has recently been accelerated due to the emergence of bigger dataset. One recent advance in this field is the use of adversarial learning to improve model learning through augmented samples. However, the use of latent features, which is feasible through adversarial learning, is not largely explored, yet. This technique may also improve the performance of affective models, as analogously demonstrated in related fields, such as computer vision. To expand this analysis, in this work, we explore the use of latent features through our proposed adversarial-based networks for valence and arousal recognition in the wild. Specifically, our models operate by aggregating several modalities to our discriminator, which is further conditioned to the extracted latent features by the generator. Our experiments on the recently released SEWA dataset suggest the progressive improvements of our results. Finally, we show our competitive results on the Affective Behavior Analysis in-the-Wild (ABAW) challenge dataset.
引用
收藏
页码:606 / 610
页数:5
相关论文
共 35 条
[1]  
[Anonymous], representation learning with deep convolutional generative
[2]  
[Anonymous], 2015, PROC 5 INT WORKSHOP
[3]  
Aspandi D, 2019, IEEE INT CONF AUTOMA, P730
[4]   Robust Facial Alignment with Internal Denoising Auto-Encoder [J].
Aspandi, Decky ;
Martinez, Oriol ;
Sukno, Federico ;
Binefa, Xavier .
2019 16TH CONFERENCE ON COMPUTER AND ROBOT VISION (CRV 2019), 2019, :143-150
[5]   StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation [J].
Choi, Yunjey ;
Choi, Minje ;
Kim, Munyoung ;
Ha, Jung-Woo ;
Kim, Sunghun ;
Choo, Jaegul .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :8789-8797
[6]   End-to-end Facial and Physiological Model for Affective Computing and Applications [J].
Comas, Joaquim ;
Aspandi, Decky ;
Binefa, Xavier .
2020 15TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2020), 2020, :93-100
[7]  
Duo S, 2010, 2010 INTERNATIONAL COLLOQUIUM ON COMPUTING, COMMUNICATION, CONTROL, AND MANAGEMENT (CCCM2010), VOL II, P360
[8]  
Eyben F., 2010, P ACM INT C MULT, P1459
[9]   The Geneva Minimalistic Acoustic Parameter Set (GeMAPS) for Voice Research and Affective Computing [J].
Eyben, Florian ;
Scherer, Klaus R. ;
Schuller, Bjoern W. ;
Sundberg, Johan ;
Andre, Elisabeth ;
Busso, Carlos ;
Devillers, Laurence Y. ;
Epps, Julien ;
Laukka, Petri ;
Narayanan, Shrikanth S. ;
Truong, Khiet P. .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2016, 7 (02) :190-202
[10]  
Gan Quan, 2015, ARXIV151106425