Evaluation of Data Inconsistency for Multi-modal Sentiment Analysis

被引:0
作者
Wang, Yufei [1 ]
Wu, Mengyue [1 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai 200000, Peoples R China
来源
MAN-MACHINE SPEECH COMMUNICATION, NCMMSC 2024 | 2025年 / 2312卷
关键词
Multi-modal Sentiment Analysis; Multi-modal Large Language Model; Data Inconsistency;
D O I
10.1007/978-981-96-1045-7_25
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Emotion semantic inconsistency is a ubiquitous challenge in multi-modal sentiment analysis (MSA). MSA involves analyzing sentiment expressed across various modalities like text, audio, and videos. Each modality may convey distinct aspects of sentiment, due to the subtle and nuanced expression of human beings, leading to inconsistency, which may hinder the prediction of artificial agents. In this work, we introduce a modality-conflicting test set and assess the performance of both traditional multi-modal sentiment analysis models and multi-modal large language models (MLLMs). Our findings reveal significant performance degradation across traditional models when confronted with semantically conflicting data and point out the drawbacks of MLLMs when handling multi-modal emotion analysis. Our research presents a new challenge and offers valuable insights for the future development of sentiment analysis systems.
引用
收藏
页码:299 / 310
页数:12
相关论文
共 50 条
  • [21] IMCN: Identifying Modal Contribution Network for Multimodal Sentiment Analysis
    Zhang, Qiongan
    Shi, Lei
    Liu, Peiyu
    Zhu, Zhenfang
    Xu, Liancheng
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4729 - 4735
  • [22] Cross-Modal Enhancement Network for Multimodal Sentiment Analysis
    Wang, Di
    Liu, Shuai
    Wang, Quan
    Tian, Yumin
    He, Lihuo
    Gao, Xinbo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 4909 - 4921
  • [23] Exploring Emotion Trends in Product Reviews: A Multi-modal Analysis with Malicious Comment Filtering and User Privacy Protection
    Chen, Biyun
    Jiang, Lin
    Pan, Xin
    Zhou, Guoquan
    Sun, Aihua
    Li, Dafang
    INFORMATION SECURITY AND CRYPTOLOGY, INSCRYPT 2023, PT I, 2024, 14526 : 379 - 396
  • [24] The Weighted Cross-Modal Attention Mechanism With Sentiment Prediction Auxiliary Task for Multimodal Sentiment Analysis
    Chen, Qiupu
    Huang, Guimin
    Wang, Yabing
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 2689 - 2695
  • [25] Hybrid cross-modal interaction learning for multimodal sentiment analysis
    Fu, Yanping
    Zhang, Zhiyuan
    Yang, Ruidi
    Yao, Cuiyou
    NEUROCOMPUTING, 2024, 571
  • [26] Dominant SIngle-Modal SUpplementary Fusion (SIMSUF) for Multimodal Sentiment Analysis
    Huang, Jian
    Ji, Yanli
    Qin, Zhen
    Yang, Yang
    Shen, Heng Tao
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 8383 - 8394
  • [27] Hybrid Contrastive Learning of Tri-Modal Representation for Multimodal Sentiment Analysis
    Mai, Sijie
    Zeng, Ying
    Zheng, Shuangjia
    Hu, Haifeng
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (03) : 2276 - 2289
  • [28] CMJRT: Cross-Modal Joint Representation Transformer for Multimodal Sentiment Analysis
    Xu, Meng
    Liang, Feifei
    Su, Xiangyi
    Fang, Cheng
    IEEE ACCESS, 2022, 10 : 131671 - 131679
  • [29] Multimodal sentiment analysis based on TCN and cross -modal interactive feedback network
    Bao, Guangbin
    Shen, Zhiming
    Liu, Chen
    Sun, Liangliang
    Chen, Shuang
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 852 - 857
  • [30] Mual: enhancing multimodal sentiment analysis with cross-modal attention and difference loss
    Deng, Yang
    Li, Yonghong
    Xian, Sidong
    Li, Laquan
    Qiu, Haiyang
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2024, 13 (03)