A Multimodal Knowledge Representation Method for Fake News Detection

被引:0
作者
Zeng, Fanhao [1 ,2 ,3 ]
Yao, Jiaxin [1 ,2 ,3 ]
Xu, Yijie [1 ,2 ,3 ]
Liu, Yanhua [1 ,2 ,3 ]
机构
[1] Fuzhou Univ, Coll Comp & Data Sci, Fuzhou, Peoples R China
[2] Minist Educ, Engn Res Ctr Big Data Intelligence, Fuzhou, Peoples R China
[3] Fuzhou Univ, Fujian Key Lab Network Comp & Intelligent Informa, Fuzhou, Peoples R China
来源
2024 4TH INTERNATIONAL CONFERENCE ON COMPUTER, CONTROL AND ROBOTICS, ICCCR 2024 | 2024年
关键词
fake news detection; multimodal representation; feature extraction; feature fusion;
D O I
10.1109/ICCCR61138.2024.10585342
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To address the key challenge encountered in fake news detection, i.e., multimodal data is difficult to be effectively semantically represented due to its intrinsic heterogeneity, this paper proposes a multimodal knowledge representation method for fake news detection. First, visual feature extraction is performed for fake news image data, the relevant images are sliced into multiple blocks, and then visual modal features are obtained by linear projection layer mapping. This simplifies the feature extraction process and reduces the computational cost, which helps to improve the fake news recognition performance. Second, to meet the actual fake news detection needs, a long text representation method based on topic words is investigated for the text data in fake news. Finally, the multimodal representation of the same fake news data is optimized by establishing a connection between two different modalities, visual and text, and inputting it into a BiLSTM-Attention based network to achieve the fusion of multimodal features. The experiment selects the same fake news data of EANN model and uses four classical classification methods to verify the effect of knowledge representation and compare it with the fusion model ViLT which is not optimized for long text. The experiment proves that the accuracy rate of fake news detection using the multimodal representation proposed in this paper is improved by 7.4% compared to the EANN model, and by 9.3% compared to the ViLT representation.
引用
收藏
页码:360 / 364
页数:5
相关论文
共 15 条
[11]  
Wang YQ, 2020, AAAI CONF ARTIF INTE, V34, P516
[12]   EANN: Event Adversarial Neural Networks for Multi-Modal Fake News Detection [J].
Wang, Yaqing ;
Ma, Fenglong ;
Jin, Zhiwei ;
Yuan, Ye ;
Xun, Guangxu ;
Jha, Kishlay ;
Su, Lu ;
Gao, Jing .
KDD'18: PROCEEDINGS OF THE 24TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2018, :849-857
[13]  
Zeng R, 2019, E-Government, P36
[14]   Multimodal Intelligence: Representation Learning, Information Fusion, and Applications [J].
Zhang, Chao ;
Yang, Zichao ;
He, Xiaodong ;
Deng, Li .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2020, 14 (03) :478-493
[15]   GRAPH ATTENTION MODEL EMBEDDED WITH MULTI-MODAL KNOWLEDGE FOR DEPRESSION DETECTION [J].
Zheng, Wenbo ;
Yan, Lan ;
Gou, Chao ;
Wang, Fei-Yue .
2020 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2020,