Multi-modal Sentiment and Emotion Joint Analysis with a Deep Attentive Multi-task Learning Model

被引:8
作者
Zhang, Yazhou [1 ]
Rong, Lu [1 ]
Li, Xiang [2 ]
Chen, Rui [1 ]
机构
[1] Zhengzhou Univ Light Ind, Software Engn Coll, Zhengzhou, Peoples R China
[2] Qilu Univ Technol, Shandong Comp Sci Ctr, Shandong Acad Sci, Natl Supercomp Ctr Jinan, Jinan, Peoples R China
来源
ADVANCES IN INFORMATION RETRIEVAL, PT I | 2022年 / 13185卷
基金
美国国家科学基金会;
关键词
Multi-modal sentiment analysis; Emotion recognition; Multi-task learning; Deep learning;
D O I
10.1007/978-3-030-99736-6_35
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Emotion is seen as the external expression of sentiment, while sentiment is the essential nature of emotion. They are tightly entangled with each other in that one helps the understanding of the other, leading to a new research topic, i.e., multi-modal sentiment and emotion joint analysis. There exists two key challenges in this field, i.e., multi-modal fusion and multi-task interaction. Most of the recent approaches treat them as two independent tasks, and fail to model the relationships between them. In this paper, we propose a novel multi-modal multi-task learning model, termed MMT, to generically address such issues. Specially, two attention mechanisms, i.e., cross-modal and cross-task attentions are designed. Cross-modal attention is proposed to model multi-modal feature fusion, while cross-task attention is to capture the interaction between sentiment analysis and emotion recognition. Finally, we empirically show that this method alleviates such problems on two benchmarking datasets, while getting better performance for the main task, i.e., sentiment analysis with the help of the secondary emotion recognition task.
引用
收藏
页码:518 / 532
页数:15
相关论文
共 30 条
[1]   A Deep Multi-task Contextual Attention Framework for Multi-modal Affect Analysis [J].
Akhtar, Md Shad ;
Chauhan, Dushyant Singh ;
Ekbal, Asif .
ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2020, 14 (03)
[2]  
Akhtar MS, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P370
[3]  
Cambria E., 2019, MULTIMODAL ANAL NEXT, P13
[4]  
Chauhan DS, 2020, 58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), P4351
[5]  
Chuang Ze-Jing, 2004, International Journal of Computational Linguistics & Chinese Language Processing, V9, P45
[6]  
Datcu D, 2015, EMOTION RECOGNITION: A PATTERN ANALYSIS APPROACH, P411
[7]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[8]  
Ghosal D, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P154
[9]   Image-text sentiment analysis via deep multimodal attentive fusion [J].
Huang, Feiran ;
Zhang, Xiaoming ;
Zhao, Zhonghua ;
Xu, Jie ;
Li, Zhoujun .
KNOWLEDGE-BASED SYSTEMS, 2019, 167 :26-37
[10]   Sentiment analysis of multimodal twitter data [J].
Kumar, Akshi ;
Garg, Geetanjali .
MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (17) :24103-24119