Puzzling Out Emotions: A Deep-Learning Approach to Multimodal Sentiment Analysis

被引:0
作者
Shrivastava, Vishal [1 ]
Richhariya, Vivek [1 ]
Richhariya, Vineet [1 ]
机构
[1] Lakshmi Narain Coll Technol, Dept Comp Sci & Engn, Bhopal, India
来源
2018 INTERNATIONAL CONFERENCE ON ADVANCED COMPUTATION AND TELECOMMUNICATION (ICACAT) | 2018年
关键词
convolutional neural networks; facial expression analysis; multi-layer perceptrons; sentiment analysis; vocal and verbal analysis; EXTRACTION; FACES;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Emotions steer both active and passive semantics of human interactions. Precise analysis of these emotions is indispensable to ensure a meaningful communication. Humans, in general, express their emotions in various forms. In order to encompass multiple dimensions of these expressions, this paper proposes a triple-layer (facial, verbal, and vocal) sentiment analysis system based on an application of deep-learning concepts. As such, in our experiment, first we separately examined facial expressions, verbal sentiments and vocal characteristics of a speaker and then mapped the individual results to perform a complete multimodal sentiment analysis. As a part of our two-stage facial expression analysis algorithm, we trained three multi-layer perceptrons using backpropagation technique to recognize a number of action units in human faces and seven single layer perceptrons each to identify one of seven basic human emotions (happiness, sadness, surprise, anger, fear, contempt or disgust, and neutral) expressed by the action units. In our vocal analysis module, we extracted important features (such as, jittering, shimmering, etc.) from sampled audio signals using standard formulae and used those features in a Bayesian Classifier to determine the type of sentiment (positive, negative, or neutral) in the voice. In the final segment of our experiment, we trained seven one dimensional convolutional neural networks to analyze verbal sentiments using the results of vocal analysis module as a bias. We were able to obtain results with as high as 91.80% (training) and 88% (testing) accuracies in our vocal and verbal analysis module; whereas, our facial expression analysis module provided results with 93.71% (training) and 92% (testing) accuracies.
引用
收藏
页数:6
相关论文
共 16 条
  • [1] Becker D., RECTIFIED LINEAR UNI
  • [2] Brown Liza, 8 BEST SOFTWARE EXTR
  • [3] Feature-point tracking by optical flow discriminates subtle differences in facial expression
    Cohn, JF
    Zlochower, AJ
    Lien, JJ
    Kanade, T
    [J]. AUTOMATIC FACE AND GESTURE RECOGNITION - THIRD IEEE INTERNATIONAL CONFERENCE PROCEEDINGS, 1998, : 396 - 401
  • [4] DARWIN, 1872, EXPRESSION EMOTIONS
  • [5] Ekman P., 1982, EMOTION HUMAN FACES
  • [6] Friesen E., 1978, FACIAL ACTION CODING, V3, P5
  • [7] Govindaraju V, 1997, IEEE SYS MAN CYBERN, P2028, DOI 10.1109/ICSMC.1997.635158
  • [8] Morin and Amy, 2015, 8 BIGGEST MYTHS LYIN
  • [9] Automatic analysis of facial expressions: The state of the art
    Pantic, M
    Rothkrantz, LJM
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2000, 22 (12) : 1424 - 1445
  • [10] Ray S., 2017, 6 Easy Steps to Learn Naive Bayes Algorithm with codes in Python and R