Multi-layer cross-modality attention fusion network for multimodal sentiment analysis

被引:1
|
作者
Yin Z. [1 ]
Du Y. [1 ]
Liu Y. [1 ]
Wang Y. [1 ]
机构
[1] Faculty of Information Technology, Beijing University of Technology, Pingleyuan 100, Beijing
基金
中国国家自然科学基金;
关键词
Cross-modality attention; Multimodal feature fusion; Multimodal sentiment analysis; Semantic alignment;
D O I
10.1007/s11042-023-17685-9
中图分类号
学科分类号
摘要
Sentiment analysis aims to detect the sentiment polarity towards the massive opinions and reviews emerging on the internet. With the increasing of multimodal information on social media, such as text, image, audio and video, multimodal sentiment analysis has attracted more attention in recent years and our work focuses on the text and image data. The previous works usually ignore the semantic alignment between the text and image, and cannot capture the interaction between them, which will affect the correct judgement for the sentiment polarity prediction. To resolve these problems, we propose a novel multimodal sentiment analysis model LXMERT-MMSA based on cross-modality attention mechanism. The single-modality feature is encoded by multi-layer Transformer encoder to achieve the deep semantic information implied in the text and image. Moreover, the cross-modality attention mechanism enables the model to fuse the text and image features effectively and achieve the rich semantic information by the alignment. It improves the ability of the model to capture the semantic relation between text and image. The evaluation metrics of accuracy and F1 score are used, and the experimental results on MVSA-multiple dataset and Twitter dataset show that our proposed model outperforms the previous SOTA model, and the ablation experimental results further prove that the model can make well use of multimodal features. © The Author(s), under exclusive licence to Springer Science+Business Media, LLC, part of Springer Nature 2024.
引用
收藏
页码:60171 / 60187
页数:16
相关论文
共 50 条
  • [1] Cross-modality Representation Interactive Learning For Multimodal Sentiment Analysis
    Huang, Jian
    Ji, Yanli
    Yang, Yang
    Shen, Heng Tao
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 426 - 434
  • [2] Multimodal Sentiment Analysis Based on Attentional Temporal Convolutional Network and Multi-Layer Feature Fusion
    Cheng, Hongju
    Yang, Zizhen
    Zhang, Xiaoqi
    Yang, Yang
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (04) : 3149 - 3163
  • [3] Self-attention Cross-modality Fusion Network for Cross-modality Person Re-identification
    Du P.
    Song Y.-H.
    Zhang X.-Y.
    Zidonghua Xuebao/Acta Automatica Sinica, 2022, 48 (06): : 1457 - 1468
  • [4] Attention fusion network for multimodal sentiment analysis
    Yuanyi Luo
    Rui Wu
    Jiafeng Liu
    Xianglong Tang
    Multimedia Tools and Applications, 2024, 83 : 8207 - 8217
  • [5] Attention fusion network for multimodal sentiment analysis
    Luo, Yuanyi
    Wu, Rui
    Liu, Jiafeng
    Tang, Xianglong
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (03) : 8207 - 8217
  • [6] Bimodal Fusion Network with Multi-Head Attention for Multimodal Sentiment Analysis
    Zhang, Rui
    Xue, Chengrong
    Qi, Qingfu
    Lin, Liyuan
    Zhang, Jing
    Zhang, Lun
    APPLIED SCIENCES-BASEL, 2023, 13 (03):
  • [7] Multimodal sentiment analysis based on multi-layer feature fusion and multi-task learning
    Cai, Yujian
    Li, Xingguang
    Zhang, Yingyu
    Li, Jinsong
    Zhu, Fazheng
    Rao, Lin
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [8] Cross-Modality Microblog Sentiment Prediction via Bi-Layer Multimodal Hypergraph Learning
    Ji, Rongrong
    Chen, Fuhai
    Cao, Liujuan
    Gao, Yue
    IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (04) : 1062 - 1075
  • [9] Cross-Modality Sentiment Analysis for Social Multimedia
    Ji, Rongrong
    Cao, Donglin
    Lin, Dazhen
    2015 1ST IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM), 2015, : 28 - 31
  • [10] A Multimodal Sentiment Analysis Method Integrating Multi-Layer Attention Interaction and Multi-Feature Enhancement
    Xie, Shengfeng
    Li, Jingwei
    INTERNATIONAL JOURNAL OF INFORMATION TECHNOLOGIES AND SYSTEMS APPROACH, 2023, 17 (01)