Multimodal Mutual Attention-Based Sentiment Analysis Framework Adapted to Complicated Contexts

被引:13
作者
He, Lijun [1 ,2 ]
Wang, Ziqing [1 ]
Wang, Liejun [3 ]
Li, Fan [1 ]
机构
[1] Xi An Jiao Tong Univ, Sch Informat & Commun Engn, Key Lab Intelligent Networks, Minist Educ, Xian 710049, Peoples R China
[2] Sichuan Digital Econ Ind Dev Res Inst, Chengdu 610036, Sichuan, Peoples R China
[3] Xinjiang Univ, Coll Informat Sci & Engn, Urumqi 830046, Peoples R China
基金
中国国家自然科学基金;
关键词
Sentiment analysis; Semantics; Tensors; Task analysis; Fuses; Neural networks; Emotion recognition; multitask framework; unimodal unique semantics; multimodal fusion; NETWORK; FUSION;
D O I
10.1109/TCSVT.2023.3276075
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Sentiment analysis has broad application prospects in the field of social opinion mining. The openness and invisibility of the internet makes users' expression styles more diverse and thus results in the blooming of complicated contexts in which different unimodal data have inconsistent sentiment tendencies. However, most sentiment analysis algorithms only focus on designing multimodal fusion methods without preserving the individual semantics of each unimodal data. To avoid misunderstandings caused by ambiguity and sarcasm in complicated contexts, we propose a multimodal mutual attention-based sentiment analysis (MMSA) framework adapted to complicated contexts, which consists of three levels of subtasks to preserve the unimodal unique semantics and enhance the common semantics, to mine the association between unique semantics and common semantics and to balance decisions from unique and common semantics. In the framework, a multiperspective and hierarchical fusion (MHF) module is developed to fully fuse multimodal data, in which different modalities are mutually constrained and the fusion order is adjusted in the next step to enhance cross-modal complementarity. To balance the data, we calculate the loss by applying different weights to positive and negative samples. The experimental results on the CH-SIMS multimodal dataset show that our method outperforms existing multimodal sentiment analysis algorithms.The code of this work is available at https://gitee.com/viviziqing/mmsacode.
引用
收藏
页码:7131 / 7143
页数:13
相关论文
共 60 条
[1]   OpenFace 2.0: Facial Behavior Analysis Toolkit [J].
Baltrusaitis, Tadas ;
Zadeh, Amir ;
Lim, Yao Chong ;
Morency, Louis-Philippe .
PROCEEDINGS 2018 13TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2018), 2018, :59-66
[2]  
Barezi EJ, 2019, 4TH WORKSHOP ON REPRESENTATION LEARNING FOR NLP (REPL4NLP-2019), P260
[3]   IEMOCAP: interactive emotional dyadic motion capture database [J].
Busso, Carlos ;
Bulut, Murtaza ;
Lee, Chi-Chun ;
Kazemzadeh, Abe ;
Mower, Emily ;
Kim, Samuel ;
Chang, Jeannette N. ;
Lee, Sungbok ;
Narayanan, Shrikanth S. .
LANGUAGE RESOURCES AND EVALUATION, 2008, 42 (04) :335-359
[4]  
Cai YT, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P2506
[5]   Benchmarking Multimodal Sentiment Analysis [J].
Cambria, Erik ;
Hazarika, Devamanyu ;
Poria, Soujanya ;
Hussain, Amir ;
Subramanyam, R. B. V. .
COMPUTATIONAL LINGUISTICS AND INTELLIGENT TEXT PROCESSING, CICLING 2017, PT II, 2018, 10762 :166-179
[6]  
Castro S, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P4619
[7]  
Chen F., 2020, P CEUR WORKSH, P82
[8]   Modeling Hierarchical Uncertainty for Multimodal Emotion Recognition in Conversation [J].
Chen, Feiyu ;
Shao, Jie ;
Zhu, Anjie ;
Ouyang, Deqiang ;
Liu, Xueliang ;
Shen, Heng Tao .
IEEE TRANSACTIONS ON CYBERNETICS, 2024, 54 (01) :187-198
[9]   Predicting Microblog Sentiments via Weakly Supervised Multimodal Deep Learning [J].
Chen, Fuhai ;
Ji, Rongrong ;
Su, Jinsong ;
Cao, Donglin ;
Gao, Yue .
IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (04) :997-1007
[10]   Verbal aggression detection on Twitter comments: convolutional neural network for short-text sentiment analysis [J].
Chen, Junyi ;
Yan, Shankai ;
Wong, Ka-Chun .
NEURAL COMPUTING & APPLICATIONS, 2020, 32 (15) :10809-10818