Multimodal Sentiment Analysis on Video Streams using Lightweight Deep Neural Networks

被引:4
|
作者
Yakaew, Atitaya [1 ]
Dailey, Matthew N. [1 ]
Racharak, Teeradaj [2 ]
机构
[1] Asian Inst Technol, Dept Informat & Commun Technol, Klongluang, Pathitimhani, Thailand
[2] Japan Adv Inst Sci & Technol, Sch Informat Sci, Nomi, Ishikawa, Japan
关键词
Deep Learning for Multimodal Real-Time Analysis; Emotion Recognition; Video Processing and Analysis; Lightweight Deep Convolutional Neural Networks; Sentiment Classification; EMOTION RECOGNITION;
D O I
10.5220/0010304404420451
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Real-time sentiment analysis on video streams involves classifying a subject's emotional expressions over time based on visual and/or audio information in the data stream. Sentiment can be analyzed using various modalities such as speech, mouth motion, and facial expression. This paper proposes a deep learning approach based on multiple modalities in which extracted features of an audiovisual data stream are fused in real time for sentiment classification. The proposed system comprises four small deep neural network models that analyze visual features and audio features concurrently. We fuse the visual and audio sentiment features into a single stream and accumulate evidence over time using an exponentially-weighted moving average to make a final prediction. Our work provides a promising solution to the problem of building real-time sentiment analysis systems that have constrained software or hardware capabilities. Experiments on the Ryerson audiovideo database of emotional speech (RAVDESS) show that deep audiovisual feature fusion yields substantial improvements over analysis of either single modality. We obtain an accuracy of 90.74%, which is better than baselines of 11.11% - 31.48% on a challenging test dataset.
引用
收藏
页码:442 / 451
页数:10
相关论文
共 50 条
  • [1] Multimodal Sentiment Analysis Using Deep Neural Networks
    Abburi, Harika
    Prasath, Rajendra
    Shrivastava, Manish
    Gangashetty, Suryakanth V.
    MINING INTELLIGENCE AND KNOWLEDGE EXPLORATION (MIKE 2016), 2017, 10089 : 58 - 65
  • [2] Sentiment Analysis of YouTube Video Comments Using Deep Neural Networks
    lassance Cunha, Alexandre Ashade
    Costa, Melissa Carvalho
    Pacheco, Marco Aurelio C.
    ARTIFICIAL INTELLIGENCEAND SOFT COMPUTING, PT I, 2019, 11508 : 561 - 570
  • [3] A Pragmatic Approach to Emoji based Multimodal Sentiment Analysis using Deep Neural Networks
    Kumar, T. Praveen
    Vardhan, B. Vishnu
    JOURNAL OF ALGEBRAIC STATISTICS, 2022, 13 (01) : 473 - 482
  • [4] Multimodal Video Sentiment Analysis Using Deep Learning Approaches, a Survey
    Abdu, Sarah A.
    Yousef, Ahmed H.
    Salem, Ashraf
    INFORMATION FUSION, 2021, 76 : 204 - 226
  • [5] Attention-based multimodal sentiment analysis and emotion recognition using deep neural networks
    Aslam, Ajwa
    Sargano, Allah Bux
    Habib, Zulfiqar
    APPLIED SOFT COMPUTING, 2023, 144
  • [6] Sentiment Analysis of Text using Deep Convolution Neural Networks
    Chachra, Anmol
    Mehndiratta, Pulkit
    Gupta, Mohit
    2017 TENTH INTERNATIONAL CONFERENCE ON CONTEMPORARY COMPUTING (IC3), 2017, : 247 - 252
  • [7] Multimodal sentiment analysis leveraging the strength of deep neural networks enhanced by the XGBoost classifier
    Chandrasekaran, Ganesh
    Dhanasekaran, S.
    Moorthy, C.
    Oli, A. Arul
    COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING, 2024,
  • [8] Multimodal Sentiment Analysis Using Deep Learning
    Sharma, Rakhee
    Le Ngoc Tan
    Sadat, Fatiha
    2018 17TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA), 2018, : 1475 - 1478
  • [9] Sentiment analysis in textual, visual and multimodal inputs using recurrent neural networks
    Jitendra V. Tembhurne
    Tausif Diwan
    Multimedia Tools and Applications, 2021, 80 : 6871 - 6910
  • [10] Sentiment analysis in textual, visual and multimodal inputs using recurrent neural networks
    Tembhurne, Jitendra V.
    Diwan, Tausif
    MULTIMEDIA TOOLS AND APPLICATIONS, 2021, 80 (05) : 6871 - 6910