Multimodal sentiment analysis based on multiple attention

被引:0
作者
Wang, Hongbin [1 ]
Ren, Chun [1 ]
Yu, Zhengtao [1 ]
机构
[1] Kunming Univ Sci & Technol, Fac Informat Engn & Automat, Kunming 650500, Yunnan, Peoples R China
关键词
Multimodal sentiment analysis; Multimodal interaction; Adaptive; Attention mechanism; TRANSFORMER;
D O I
10.1016/j.engappai.2024.109731
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of the Internet makes various types of data widely appear on various social platforms, multimodal data provides anew perspective for sentiment analysis. Although the data types are different, there are information expressing the same sentiment. The existing researches on extracting those information are static, and this means that there is a problem of extracting common information in a fixed amount. Therefore, to address this problem, we proposes a method named multimodal sentiment analysis based on multiple attention(MAMSA). Firstly, this method utilized the adaptive attention interaction module to dynamically determine the amount of information contributed by text and image features in multimodal fusion, and multimodal common representations are extracted through cross modal attention to improve the performance of each modal feature representation. Secondly, using sentiment information as a guide to extract text and image features related to sentiment. Finally, using hierarchical manner to fully learning the internal correlations between sentiment-text association representation, sentiment-image association representation, and multimodal common information to improve the performance of the model. We conducted extensive experiments using two public multimodal datasets, and the experimental results validated the availability of the proposed method.
引用
收藏
页数:10
相关论文
共 37 条
  • [1] Joint multimodal sentiment analysis based on information relevance
    Chen, Danlei
    Su, Wang
    Wu, Peng
    Hua, Bolin
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (02)
  • [2] MISA: Modality-Invariant and -Specific Representations for Multimodal Sentiment Analysis
    Hazarika, Devamanyu
    Zimmermann, Roger
    Poria, Soujanya
    [J]. MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1122 - 1131
  • [3] Dynamic Invariant-Specific Representation Fusion Network for Multimodal Sentiment Analysis
    He, Jing
    Yanga, Haonan
    Zhang, Changfan
    Chen, Hongrun
    Xua, Yifu
    [J]. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [4] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [5] TeFNA: Text-centered fusion network with crossmodal attention for multimodal sentiment analysis
    Huang, Changqin
    Zhang, Junling
    Wu, Xuemei
    Wang, Yi
    Li, Ming
    Huang, Xiaodi
    [J]. KNOWLEDGE-BASED SYSTEMS, 2023, 269
  • [6] Huang LZ, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P3444
  • [7] Kim Y., 2014, P 2014 C EMP METH NA, P1746
  • [8] Kingma D.P., 2014, arXiv
  • [9] Lai SW, 2015, AAAI CONF ARTIF INTE, P2267
  • [10] Multi-Label Multimodal Emotion Recognition With Transformer-Based Fusion and Emotion-Level Representation Learning
    Le, Hoai-Duy
    Lee, Guee-Sang
    Kim, Soo-Hyung
    Kim, Seungwon
    Yang, Hyung-Jeong
    [J]. IEEE ACCESS, 2023, 11 : 14742 - 14751