Bimodal sentiment analysis in social media: a one-shot learning approach

被引:0
作者
Pakdaman, Zahra [1 ]
Koochari, Abbas [1 ]
Sharifi, Arash [1 ]
机构
[1] Islamic Azad Univ, Dept Comp Engn, Sci & Res Branch, Tehran, Iran
关键词
Visual Sentiment Analysis; Sentiment Image Captioning; One-Shot Learning; Relation Network; Semantic Text Similarity;
D O I
10.1007/s11042-024-18748-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the rapid growth of social media (forums, blogs, social networks, etc.) entities like polling organizations, advertising companies, as well as democratic organizations are seeking to discover the opinions of their audiences more than ever. Analyzing such information helps them to perform operations like production monitoring, market forecasting, and civil society demands with high accuracy. Besides these entities, individuals/customers also choose from other users' feedback. The most basic processing that takes place in the field of sentiment analysis is to classify a text into a positive or negative class. Yet, given to wide usage of image-based social media, there have been limited attempts to deduce emotions from images or videos. In this paper, a bimodal sentiment analysis is presented by combining visual and textual information. In the proposed method, the image is sentimentally captioned by an encoder-decoder network, first. To achieve this goal, a sentiment image captioning method has been used to generate positive, negative, or neutral captions for each input image. Then, the generated captions and the user's published text are sent to the relation network and then the final sentiment label is predicted. In this one-shot learning approach, only one image is used to predict a sentiment, without any training. Results show the proposed method can compete with the State-of-the-Art methods in the field of visual sentiment analysis. In this respect, the proposed method achieves 70% accuracy on TWITTER and 75% accuracy on MVSA datasets, which can be regarded as a champion considering that this method is a one-shot approach.
引用
收藏
页码:80351 / 80372
页数:22
相关论文
共 70 条
  • [1] AndrewShin Y.U., 2016, P BRIT MACHINE VISIO
  • [2] Multimodal sentiment and emotion recognition in hyperbolic space
    Arano, Keith April
    Orsenigo, Carlotta
    Soto, Mauricio
    Vercellis, Carlo
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2021, 184
  • [3] Asakawa T, 2020, ASIAPAC SIGN INFO PR, P1081
  • [4] Ayetiran E., 2022, Attention-based aspect sentiment classification using enhanced learning through CNN-BiLSTM network, DOI [10.1016/j.knosys.2022.109409, DOI 10.1016/J.KNOSYS.2022.109409]
  • [5] Borth Damian, 2013, P ACM INT C MULT, P2, DOI DOI 10.1145/2502081.2502282
  • [6] Interactions Guided Generative Adversarial Network for unsupervised image captioning
    Cao, Shan
    An, Gaoyun
    Zheng, Zhenxing
    Ruan, Qiuqi
    [J]. NEUROCOMPUTING, 2020, 417 : 419 - 431
  • [7] Cer D, 2018, Arxiv, DOI arXiv:1803.11175
  • [8] Chen M, 2018, P 19 ACM INT C MULTI, P163, DOI [10.48550/arXiv.1802.00924, DOI 10.48550/ARXIV.1802.00924]
  • [9] Chen TL, 2018, Arxiv, DOI [arXiv:1807.03871, 10.48550/arXiv.1807.03871]
  • [10] Deng CR, 2020, Arxiv, DOI arXiv:2007.09580