A Co-Memory Network for Multimodal Sentiment Analysis

被引:119
作者
Xu, Nan [1 ]
Mao, Wenji [1 ]
Chen, Guandan [1 ]
机构
[1] Chinese Acad Sci, Univ Chinese Acad Sci, Inst Automat, Beijing, Peoples R China
来源
ACM/SIGIR PROCEEDINGS 2018 | 2018年
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Multimodal Sentiment Analysis; Co-Memory Network; Attention;
D O I
10.1145/3209978.3210093
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the rapid increase of diversity and modality of data in usergenerated contents, sentiment analysis as a core area of social media analytics has gone beyond traditional text-based analysis. Multimodal sentiment analysis has become an important research topic in recent years. Most of the existing work on multimodal sentiment analysis extracts features from image and text separately, and directly combine them to train a classifier. As visual and textual information in multimodal data can mutually reinforce and complement each other in analyzing the sentiment of people, previous research all ignores this mutual influence between image and text. To fill this gap, in this paper, we consider the interrelation of visual and textual information, and propose a novel co-memory network to iteratively model the interactions between visual contents and textual words for multimodal sentiment analysis. Experimental results on two public multimodal sentiment datasets demonstrate the effectiveness of our proposed model compared to the state-of-the-art methods.
引用
收藏
页码:929 / 932
页数:4
相关论文
共 13 条
[1]  
[Anonymous], 2016, NAT METHODS, DOI DOI 10.1038/nmeth.3707
[2]   A multimodal feature learning approach for sentiment analysis of social network multimedia [J].
Baecchi, Claudio ;
Uricchio, Tiberio ;
Bertini, Marco ;
Del Bimbo, Alberto .
MULTIMEDIA TOOLS AND APPLICATIONS, 2016, 75 (05) :2507-2525
[3]  
Borth D., 2013, Proceedings of the 21st ACM international conference on multimedia, P223
[4]   Convolutional Neural Networks for Multimedia Sentiment Analysis [J].
Cai, Guoyong ;
Xia, Binbin .
NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2015, 2015, 9362 :159-167
[5]  
Li Z, 2017, PROCEEDINGS OF THE TWENTY-SIXTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, P2237
[6]  
Niu T., 2016, P 22 INT C MULTIMED, V9517, P15, DOI [DOI 10.1007/978-3-319-27674-82, 10.1007/978-3-319-27674-8_2]
[7]  
Pennington J., 2014, 2014 C EMP METH NAT, P43
[8]   A survey of multimodal sentiment analysis [J].
Soleymani, Mohammad ;
Garcia, David ;
Jou, Brendan ;
Schuller, Bjoern ;
Chang, Shih-Fu ;
Pantic, Maja .
IMAGE AND VISION COMPUTING, 2017, 65 :3-14
[9]   Rethinking the Inception Architecture for Computer Vision [J].
Szegedy, Christian ;
Vanhoucke, Vincent ;
Ioffe, Sergey ;
Shlens, Jon ;
Wojna, Zbigniew .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :2818-2826
[10]   MultiSentiNet: A Deep Semantic Network for Multimodal Sentiment Analysis [J].
Xu, Nan ;
Mao, Wenji .
CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, :2399-2402