Multimodal sentiment analysis based on fusion methods: A survey

被引:138
作者
Zhu, Linan [1 ]
Zhu, Zhechao [1 ]
Zhang, Chenwei [2 ]
Xu, Yifei [1 ]
Kong, Xiangjie [1 ]
机构
[1] Zhejiang Univ Technol, Coll Comp Sci & Technol, Hangzhou, Peoples R China
[2] Univ Hong Kong, Sch Fac Educ, Hong Kong, Peoples R China
基金
中国国家自然科学基金;
关键词
Multimodal data; Sentiment analysis; Feature extraction; Fusion methods; NETWORK;
D O I
10.1016/j.inffus.2023.02.028
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Sentiment analysis is an emerging technology that aims to explore people's attitudes toward an entity. It can be applied in a variety of different fields and scenarios, such as product review analysis, public opinion analysis, psychological disease analysis, and risk assessment analysis. Traditional sentiment analysis only includes the text modality and extracts sentiment information by inferring the semantic relationship within sentences. However, some special expressions, such as irony and exaggeration, are difficult to detect via text alone. Multimodal sentiment analysis contains rich visual and acoustic information in addition to text, and uses fusion analysis to more accurately infer the implied sentiment polarity (positive, neutral, negative). The main challenge in multimodal sentiment analysis is the integration of cross-modal sentiment information, so we focus on introducing the framework and characteristics of different fusion methods. In addition, this article discusses the development status of multimodal sentiment analysis, popular datasets, feature extraction algorithms, application areas, and existing challenges. It is hoped that our work can help researchers understand the current state of research in the field of multimodal sentiment analysis, and be inspired by the useful insights provided in the article to develop effective models.
引用
收藏
页码:306 / 325
页数:20
相关论文
共 107 条
[11]   Recognition of Action Units in the Wild with Deep Nets and a New Global-Local Loss [J].
Benitez-Quiroz, C. Fabian ;
Wang, Yan ;
Martinez, Aleix M. .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :3990-3999
[12]  
Borth D., 2013, P ACM INT C MULT, P223, DOI 10.1145/2502081.2502282
[13]   Guest Editorial: Big Social Data Analysis [J].
Cambria, Erik ;
Wang, Haixun ;
White, Bebo .
KNOWLEDGE-BASED SYSTEMS, 2014, 69 :1-2
[14]  
Castro S, 2019, 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), P4619
[15]   Multimodal sentimental analysis for social media applications: A comprehensive review [J].
Chandrasekaran, Ganesh ;
Nguyen, Tu N. ;
Hemanth, D. Jude .
WILEY INTERDISCIPLINARY REVIEWS-DATA MINING AND KNOWLEDGE DISCOVERY, 2021, 11 (05)
[16]  
Chauhan DS, 2019, 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019), P5647
[17]  
Chen MH, 2017, PROCEEDINGS OF THE 19TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, ICMI 2017, P163, DOI 10.1145/3136755.3136801
[18]  
Cho K., 2014, ARXIV, DOI [10.3115/v1/W14-4012, DOI 10.3115/V1/W14-4012]
[19]   SUPPORT-VECTOR NETWORKS [J].
CORTES, C ;
VAPNIK, V .
MACHINE LEARNING, 1995, 20 (03) :273-297
[20]   A Review and Meta-Analysis of Multimodal Affect Detection Systems [J].
D'Mello, Sidney K. ;
Kory, Jacqueline .
ACM COMPUTING SURVEYS, 2015, 47 (03)