An Out-of-Distribution Attack Resistance Approach to Emotion Categorization

被引:5
作者
Shehu H.A. [1 ]
Browne W.N. [2 ]
Eisenbarth H. [3 ]
机构
[1] The School of Engineering and Computer Science, Victoria University of Wellington, Wellington
[2] The School of Electrical Engineering and Robotics, Queensland University of Technology, Brisbane
[3] The School of Psychology, Victoria University of Wellington, Wellington
来源
IEEE Transactions on Artificial Intelligence | 2021年 / 2卷 / 06期
关键词
Attack; cross-database; emotion categorization; emotion recognition; facial expression; facial landmarks; out-of-distribution;
D O I
10.1109/TAI.2021.3105371
中图分类号
学科分类号
摘要
Deep neural networks are a powerful model for feature extraction. They produce features that enable state-of-the-art performance on many tasks, including emotion categorization. However, their homogeneous representation of knowledge has made them prone to attacks, i.e., small modification in train or test data to mislead the models. Emotion categorization can usually be performed to be either in-distribution (train and test with the same dataset) or out-of-distribution (train on one or more dataset(s) and test on a different dataset). Our already developed landmark-based technique, which is robust for in-distribution improvement against attacks in emotion categorization, could translate to out-of-distribution classification problems. This is important as different databases might have different variations such as in color or level of expressiveness of emotion. We compared the landmark-based method with four state-of-the-art deep models (EfficientNetB0, InceptionV3, ResNet50, and VGG19), as well as emotion categorization tools (i.e., Python Facial Expression Analysis Toolbox and the Microsoft Azure face application programming interface) by performing a cross-database experiment across six commonly used databases, i.e., extended Cohn-Kanade, Japanese female facial expression, Karolinska directed emotional faces, National Institute of Mental Health Child Emotional Faces Picture Set, real-world affective faces, and psychological image collection at Stirling databases. The landmark-based method has achieved a significantly higher accuracy, achieving an average of 47.44% compared with most of the deep networks (< 36%) and the emotion categorization tools (<37%) with considerably less execution time. This highlights that out-of-distribution emotion categorization is a much harder task due to detecting underlying emotional cues than emotion categorization in-distribution where superficial patterns are detected to > 97% accuracy. Impact Statement-Recognising emotions from people's faces has real-world applications for computer-based perception as it is often vital for interpersonal communication. Emotion recognition tasks nowadays are addressed using deep learning models that model colour distribution so classify images rather than emotion. This homogeneous knowledge representation is in contrast to emotion categorization, which is hypothesised as more heterogeneous landmark-based. This is investigated through out-of-distribution emotion categorization problems, where the test samples are drawn from a different dataset to training images. Our landmark-based method achieves a significantly higher classification performance (on average) compared with four state-of-the-art deep networks (EfficientNetB0, InceptionV3, ResNet50 and VGG19), as well as other emotion categorization tools such as Py-Feat and the Azure Face API. We conclude that this improved generalization is relevant for future developments of emotion categorization tools. © 2021 IEEE.
引用
收藏
页码:564 / 573
页数:9
相关论文
共 50 条
  • [1] Ren F., Bao Y., A review on human-computer interaction and intelligent robots, Int. J. Inf. Technol. Decis. Making, 19, 1, pp. 5-47, (2020)
  • [2] Li S., Deng W., Deep facial expression recognition: A survey, IEEE Trans. Affective Comput., (2020)
  • [3] Nicholson-Smith C., Mehrabi V., Atashzar S.F., Patel R.V., A multifunctional lower- and upper-limb stroke rehabilitation robot, IEEE Trans. Med. Robot. Bionics, 2, 4, pp. 549-552, (2020)
  • [4] Tang Y., Zhang X., Hu X., Wang S., Wang H., Facial expression recognition using frequency neural network, IEEE Trans. Image Process., 30, pp. 444-457, (2020)
  • [5] Khnlenz K., Radig B., Mayer C., Sosnowski S., Towards robotic facial mimicry: System development and evaluation, Proc. 19th Int. Symp. Robot Hum. Interactive Commun., pp. 198-203, (2011)
  • [6] Kim S., Kim H., Deep explanation model for facial expression recognition through facial action coding unit, Proc. IEEE Int. Conf. Big Data Smart Comput., pp. 1-4, (2019)
  • [7] Zhang S., Pan X., Cui Y., Zhao X., Liu L., Learning affective video features for facial expression recognition via hybrid deep learning, IEEE Access, 7, pp. 32297-32304, (2019)
  • [8] Valstar M.F., Mehu M., Jiang B., Pantic M., Scherer K., Metaanalysis of the first facial expression recognition challenge, IEEE Trans. Syst., Man, Cybern., Part B. (Cybern.), 42, 4, pp. 966-979, (2012)
  • [9] Intahchomphoo C., Gundersen O., Artificial intelligence and race: A systematic review, Legal Inf. Manage., 2, 20, pp. 74-84, (2020)
  • [10] Shehu H.A., Browne W., Eisenbarth H., An adversarial attacks resistance-based approach to emotion recognition from images using facial landmarks, Proc. 29th IEEE Int. Conf. Robot Hum. Interactive Commun., pp. 1307-1314, (2020)