A text guided multi-task learning network for multimodal sentiment analysis

被引:6
|
作者
Luo, Yuanyi [1 ]
Wu, Rui [1 ]
Liu, Jiafeng [1 ]
Tang, Xianglong [1 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin 150001, Peoples R China
关键词
Multimodal sentiment analysis; Representation learning; Multi-task learning; FUSION;
D O I
10.1016/j.neucom.2023.126836
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal Sentiment Analysis (MSA) is an active area of research that leverages multimodal signals for affective understanding of user-generated videos. Existing research tends to develop sophisticated fusion techniques to fuse unimodal representations into multimodal representation and treat MSA as a single prediction task. However, we find that the text modality with the pre-trained model (BERT) learn more semantic information and dominates the training in multimodal models, inhibiting the learning of other modalities. Besides, the classification ability of each modality is also suppressed by single-task learning. In this paper, We propose a text guided multi-task learning network to enhance the semantic information of non-text modalities and improve the learning ability of unimodal networks. We conducted experiments on multimodal sentiment analysis datasets, CMU-MOSI, CMU-MOSEI, and CH-SIMS. The results show that our method outperforms the current SOTA method.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Multimodal Sentiment Recognition With Multi-Task Learning
    Zhang, Sun
    Yin, Chunyong
    Yin, Zhichao
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2023, 7 (01): : 200 - 209
  • [2] A Dual-branch Enhanced Multi-task Learning Network for Multimodal Sentiment Analysis
    Geng, Wenxiu
    Li, Xiangxian
    Bian, Yulong
    PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 481 - 489
  • [3] Multimodal Sentiment Analysis With Two-Phase Multi-Task Learning
    Yang, Bo
    Wu, Lijun
    Zhu, Jinhua
    Shao, Bo
    Lin, Xiaola
    Liu, Tie-Yan
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 2015 - 2024
  • [4] Multi-Task Momentum Distillation for Multimodal Sentiment Analysis
    Lin, Ronghao
    Hu, Haifeng
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (02) : 549 - 565
  • [5] Multi-level Multi-task representation learning with adaptive fusion for multimodal sentiment analysis
    Chuanbo Zhu
    Min Chen
    Haomin Li
    Sheng Zhang
    Han Liang
    Chao Sun
    Yifan Liu
    Jincai Chen
    Neural Computing and Applications, 2025, 37 (3) : 1491 - 1508
  • [6] Multimodal sentiment analysis based on multi-layer feature fusion and multi-task learning
    Cai, Yujian
    Li, Xingguang
    Zhang, Yingyu
    Li, Jinsong
    Zhu, Fazheng
    Rao, Lin
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [7] Multi-task learning and mutual information maximization with crossmodal transformer for multimodal sentiment analysis
    Shi, Yang
    Cai, Jinglang
    Liao, Lei
    JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2024, : 1 - 19
  • [8] Multi-task disagreement-reducing multimodal sentiment fusion network
    Wang, Zijun
    Jiang, Naicheng
    Chao, Xinyue
    Sun, Bin
    IMAGE AND VISION COMPUTING, 2024, 149
  • [9] MCapsNet: Capsule Network for Text with Multi-Task Learning
    Xiao, Liqiang
    Zhang, Honglun
    Chen, Wenqing
    Wang, Yongkun
    Jin, Yaohui
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4565 - 4574
  • [10] Improving sentiment analysis with multi-task learning of negation
    Barnes, Jeremy
    Velldal, Erik
    Ovrelid, Lilja
    NATURAL LANGUAGE ENGINEERING, 2021, 27 (02) : 249 - 269