CubeMLP: A MLP-based Model for Multimodal Sentiment Analysis and Depression Estimation

被引:55
|
作者
Sun, Hao [1 ]
Wang, Hongyi [1 ]
Liu, Jiaqing [2 ]
Chen, Yen-Wei [2 ]
Lin, Lanfen [1 ]
机构
[1] Zhejiang Univ, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[2] Ritsumeikan Univ, Coll Informat Sci & Engn, Kusatsu, Shiga, Japan
来源
PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022 | 2022年
关键词
multimodal processing; multimodal fusion; multimodal interaction; multimedia; MLP; sentiment analysis; depression detection;
D O I
10.1145/3503161.3548025
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Multimodal sentiment analysis and depression estimation are two important research topics that aim to predict human mental states using multimodal data. Previous research has focused on developing effective fusion strategies for exchanging and integrating mind-related information from different modalities. Some MLP-based techniques have recently achieved considerable success in a variety of computer vision tasks. Inspired by this, we explore multimodal approaches with a feature-mixing perspective in this study. To this end, we introduce CubeMLP, a multimodal feature processing framework based entirely on MLP. CubeMLP consists of three independent MLP units, each of which has two affine transformations. CubeMLP accepts all relevant modality features as input and mixes them across three axes. After extracting the characteristics using CubeMLP, the mixed multimodal features are flattened for task predictions. Our experiments are conducted on sentiment analysis datasets: CMU-MOSI and CMU-MOSEI, and depression estimation dataset: AVEC2019. The results show that CubeMLP can achieve state-of-the-art performance with a much lower computing cost.
引用
收藏
页码:3722 / 3729
页数:8
相关论文
共 50 条
  • [21] Multimodal Sentiment Analysis Based on Bidirectional Mask Attention Mechanism
    Zhang Y.
    Zhang H.
    Liu Y.
    Liang K.
    Wang Y.
    Data Analysis and Knowledge Discovery, 2023, 7 (04) : 46 - 55
  • [22] Multimodal Sentiment Analysis Based on Expert Mixing of Subtask Representations
    Lei, Ling
    He, Wangjun
    Zheng, Qiuyan
    Zhu, Bing
    IEEE ACCESS, 2025, 13 : 44278 - 44287
  • [23] A STUDENT SENTIMENT ANALYSIS METHOD BASED ON MULTIMODAL DEEP LEARNING
    Kong, Lidan
    Yao, Jian
    Shen, Jinsong
    Gu, Yi
    JOURNAL OF MECHANICS IN MEDICINE AND BIOLOGY, 2024, 24 (09)
  • [24] A Multimodal Sentiment Analysis Method Based on Fuzzy Attention Fusion
    Zhi, Yuxing
    Li, Junhuai
    Wang, Huaijun
    Chen, Jing
    Wei, Wei
    IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2024, 32 (10) : 5886 - 5898
  • [25] Multimodal Consistency-Based Teacher for Semi-Supervised Multimodal Sentiment Analysis
    Yuan, Ziqi
    Fang, Jingliang
    Xu, Hua
    Gao, Kai
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3669 - 3683
  • [26] Multimodal Mutual Attention-Based Sentiment Analysis Framework Adapted to Complicated Contexts
    He, Lijun
    Wang, Ziqing
    Wang, Liejun
    Li, Fan
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) : 7131 - 7143
  • [27] Self-adaptive attention fusion for multimodal aspect-based sentiment analysis
    Wang, Ziyue
    Guo, Junjun
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2024, 21 (01) : 1305 - 1320
  • [28] Adversarial attack evaluation and defense method for multimodal sentiment analysis model
    Fan F.
    Nie X.
    Deng X.
    Liu S.
    Huazhong Keji Daxue Xuebao (Ziran Kexue Ban)/Journal of Huazhong University of Science and Technology (Natural Science Edition), 2023, 51 (02): : 19 - 24
  • [29] AB-GRU: An attention-based bidirectional GRU model for multimodal sentiment fusion and analysis
    Wu, Jun
    Zheng, Xinli
    Wang, Jiangpeng
    Wu, Junwei
    Wang, Ji
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (10) : 18523 - 18544
  • [30] Tree-Based Mix-Order Polynomial Fusion Network for Multimodal Sentiment Analysis
    Tang, Jiajia
    Hou, Ming
    Jin, Xuanyu
    Zhang, Jianhai
    Zhao, Qibin
    Kong, Wanzeng
    SYSTEMS, 2023, 11 (01):