Sentiment analysis method of consumer reviews based on multi-modal feature mining

被引:0
作者
You, Jing [1 ,2 ]
Zhong, Jiamin [3 ]
Kong, Jing [1 ]
Peng, Lihua [4 ]
机构
[1] School of Economics and Management, Guangdong Open University (Guangdong Polytechnic Institute), Guangdong, Guangzhou
[2] Agricultural Industry and Digital Economy Research Center of Guangdong Open University (Guangdong Polytechnic Institute), Guangdong, Guangzhou
[3] Zhongshan Xiaoji Technology Co., Ltd, Guangdong, Zhongshan
[4] School of Law and Administration, Guangdong Open University (Guangdong Polytechnic Institute), Guangdong, Guangzhou
来源
International Journal of Cognitive Computing in Engineering | 2025年 / 6卷
关键词
Cross attention; Mask graph transformer; Sentiment analysis; VIT and BGE;
D O I
10.1016/j.ijcce.2024.12.001
中图分类号
学科分类号
摘要
Traditional sentiment analysis methods primarily rely on textual data. However, in real-world applications, product reviews often contain multimodal information including images, videos, and audio. This multimodal data is crucial for accurately understanding consumer sentiment trends. Therefore, this article proposes a novel product review sentiment analysis method that leverages multimodal feature mining. Firstly, a pre-trained vision transformer (VIT) and baidu general embedding (BGE) models are utilized to encode image and text features. Then, internal correlation mining is performed on image and text features through cross-attention. Next, the image and text features are aligned using a mask graph transformer network. The final encoding is obtained using multi-layer perceptron. Lastly, cosine similarity is computed between this encoding and each sentiment aspect of BGE encoding to determine corresponding sentiment scores. Simulations were conducted on a multi-modal Multi-ZOL dataset to compare the proposed method with several state-of-the-art techniques. The obtained results validated the superior performance of the proposed method. © 2024
引用
收藏
页码:143 / 151
页数:8
相关论文
共 47 条
[1]  
Abdu S.A., Yousef A.H., Salem A., Multimodal video sentiment analysis using deep learning approaches, a survey, Information Fusion, 76, pp. 204-226, (2021)
[2]  
Al-Dabet S., Tedmori S., Mohammad A.S., Enhancing Arabic aspect-based sentiment analysis using deep learning models, Computer Speech & Language, 69, (2021)
[3]  
Araque O., Corcuera-Platas I., Sanchez-Rada J.F., Iglesias C.A., Enhancing deep learning sentiment analysis with ensemble techniques in social applications, Expert Systems with Applications, 77, pp. 236-246, (2017)
[4]  
Birjali M., Kasri M., Beni-Hssane A., A comprehensive survey on sentiment analysis: Approaches, challenges and trends, Knowledge-Based Systems, 226, (2021)
[5]  
Chen D., Su W., Wu P., Hua B., Joint multimodal sentiment analysis based on information relevance, Information Processing & Management, 60, 2, (2023)
[6]  
Chen P., Sun Z., Bing L., Yang W., Recurrent attention network on memory for aspect sentiment analysis, Proceedings of the 2017 conference on empirical methods in natural language processing, pp. 452-461, (2017)
[7]  
Da Silva N.F.F., Coletta L.F., Hruschka E.R., Hruschka E.R., Using unsupervised information to improve semi-supervised tweet sentiment classification, Information Sciences, 355, pp. 348-365, (2016)
[8]  
Demircan M., Seller A., Abut F., Akay M.F., Developing Turkish sentiment analysis models using machine learning and e-commerce data, International Journal of Cognitive Computing in Engineering, 2, pp. 202-207, (2021)
[9]  
Devlin J., Chang M.W., Lee K., Toutanova K., Bert: Pre-training of deep bidirectional transformers for language understanding, (2018)
[10]  
Gong X., Ren J., Zeng L., Xing R., How KOLs influence consumer purchase intention in short video platforms: Mediating effects of emotional response and virtual touch, International Journal of Information Systems in the Service Sector (IJISSS), 14, 1, pp. 1-23, (2022)