Towards Robust Multimodal Sentiment Analysis Under Uncertain Signal Missing

被引:14
作者
Li, Mingcheng [1 ]
Yang, Dingkang [1 ]
Zhang, Lihua [1 ]
机构
[1] Fudan Univ, Acad Engn & Technol, Shanghai 200433, Peoples R China
基金
国家重点研发计划;
关键词
Semantics; Feature extraction; Transformers; Sentiment analysis; Visualization; Training; Feeds; Multimodal sentiment analysis; crossmodal interaction; knowledge distillation; uncertain signal missing;
D O I
10.1109/LSP.2023.3324552
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Multimodal Sentiment Analysis (MSA) has attracted widespread research attention recently. Most MSA studies are based on the assumption of signal completeness. However, many inevitable factors in real applications lead to uncertain signal missing, causing significant degradation of model performance. To this end, we propose a Robust multimodal Missing Signal Framework (RMSF) to handle the problem of uncertain signal missing for MSA tasks and can be generalized to other multimodal patterns. Specifically, a hierarchical crossmodal interaction module in RMSF exploits potential complementary semantics among modalities via coarse- and fine-grained crossmodal attention. Furthermore, we design an adaptive feature refinement module to enhance the beneficial semantics of modalities and filter redundant features. Finally, we propose a knowledge-integrated self-distillation module that enables dynamic knowledge integration and bidirectional knowledge transfer within a single network to precisely reconstruct missing semantics. Comprehensive experiments are conducted on two datasets, indicating that RMSF significantly improves MSA performance under both uncertain missing-signal and complete-signal cases.
引用
收藏
页码:1497 / 1501
页数:5
相关论文
共 40 条
  • [1] Anil Rohan, 2018, INT C LEARN REPR
  • [2] Baltrusaitis T, 2016, IEEE WINT CONF APPL
  • [3] Dealing with Missing Modalities in the Visual Question Answer-Difference Prediction Task through Knowledge Distillation
    Cho, Jae Won
    Kim, Dong-Jin
    Choi, Jinsoo
    Jung, Yunjae
    Kweon, In So
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 1592 - 1601
  • [4] Degottex G, 2014, INT CONF ACOUST SPEE, DOI 10.1109/ICASSP.2014.6853739
  • [5] Semi-supervised Deep Generative Modelling of Incomplete Multi-Modality Emotional Data
    Du, Changde
    Du, Changying
    Wang, Hao
    Li, Jinpeng
    Zheng, Wei-Long
    Lu, Bao-Liang
    He, Huiguang
    [J]. PROCEEDINGS OF THE 2018 ACM MULTIMEDIA CONFERENCE (MM'18), 2018, : 108 - 116
  • [6] LEARNING ASSOCIATIVE REPRESENTATION FOR FACIAL EXPRESSION RECOGNITION
    Du, Yangtao
    Yang, Dingkang
    Zhai, Peng
    Li, Mingchen
    Zhang, Lihua
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 889 - 893
  • [7] Han W, 2021, 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), P9180
  • [8] MISA: Modality-Invariant and -Specific Representations for Multimodal Sentiment Analysis
    Hazarika, Devamanyu
    Zimmermann, Roger
    Poria, Soujanya
    [J]. MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 1122 - 1131
  • [9] Language Reinforced Superposition Multimodal Fusion for Sentiment Analysis
    He, Jiaxuan
    Hu, Haifeng
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 1347 - 1351
  • [10] MF-BERT: Multimodal Fusion in Pre-Trained BERT for Sentiment Analysis
    He, Jiaxuan
    Hu, Haifeng
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 454 - 458