A Multimodal Knowledge Representation Method for Fake News Detection

被引:0
作者
Zeng, Fanhao [1 ,2 ,3 ]
Yao, Jiaxin [1 ,2 ,3 ]
Xu, Yijie [1 ,2 ,3 ]
Liu, Yanhua [1 ,2 ,3 ]
机构
[1] Fuzhou Univ, Coll Comp & Data Sci, Fuzhou, Peoples R China
[2] Minist Educ, Engn Res Ctr Big Data Intelligence, Fuzhou, Peoples R China
[3] Fuzhou Univ, Fujian Key Lab Network Comp & Intelligent Informa, Fuzhou, Peoples R China
来源
2024 4TH INTERNATIONAL CONFERENCE ON COMPUTER, CONTROL AND ROBOTICS, ICCCR 2024 | 2024年
关键词
fake news detection; multimodal representation; feature extraction; feature fusion;
D O I
10.1109/ICCCR61138.2024.10585342
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
To address the key challenge encountered in fake news detection, i.e., multimodal data is difficult to be effectively semantically represented due to its intrinsic heterogeneity, this paper proposes a multimodal knowledge representation method for fake news detection. First, visual feature extraction is performed for fake news image data, the relevant images are sliced into multiple blocks, and then visual modal features are obtained by linear projection layer mapping. This simplifies the feature extraction process and reduces the computational cost, which helps to improve the fake news recognition performance. Second, to meet the actual fake news detection needs, a long text representation method based on topic words is investigated for the text data in fake news. Finally, the multimodal representation of the same fake news data is optimized by establishing a connection between two different modalities, visual and text, and inputting it into a BiLSTM-Attention based network to achieve the fusion of multimodal features. The experiment selects the same fake news data of EANN model and uses four classical classification methods to verify the effect of knowledge representation and compare it with the fusion model ViLT which is not optimized for long text. The experiment proves that the accuracy rate of fake news detection using the multimodal representation proposed in this paper is improved by 7.4% compared to the EANN model, and by 9.3% compared to the ViLT representation.
引用
收藏
页码:360 / 364
页数:5
相关论文
共 15 条
[1]  
Breiman L., 2001, MACH LEARN, V45, P5
[2]   UNITER: UNiversal Image-TExt Representation Learning [J].
Chen, Yen-Chun ;
Li, Linjie ;
Yu, Licheng ;
El Kholy, Ahmed ;
Ahmed, Faisal ;
Gan, Zhe ;
Cheng, Yu ;
Liu, Jingjing .
COMPUTER VISION - ECCV 2020, PT XXX, 2020, 12375 :104-120
[3]   SUPPORT-VECTOR NETWORKS [J].
CORTES, C ;
VAPNIK, V .
MACHINE LEARNING, 1995, 20 (03) :273-297
[4]   NEAREST NEIGHBOR PATTERN CLASSIFICATION [J].
COVER, TM ;
HART, PE .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1967, 13 (01) :21-+
[5]  
Cui YM, 2020, Arxiv, DOI [arXiv:2004.13922, DOI 10.48550/ARXIV.2004.13922]
[6]  
Huang ZC, 2020, Arxiv, DOI [arXiv:2004.00849, 10.48550/arXiv.2004.00849]
[7]  
Keybert Grootendorst M., 2020, Minimal keyword extraction with bert, DOI [10.5281/zenodo.446126, DOI 10.5281/ZENODO.446126]
[8]  
Kim W, 2021, PR MACH LEARN RES, V139
[9]   Knowledge-aware Multi-modal Adaptive Graph Convolutional Networks for Fake News Detection [J].
Qian, Shengsheng ;
Hu, Jun ;
Fang, Quan ;
Xu, Changsheng .
ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2021, 17 (03)
[10]  
Quinlan J. R., 1986, Machine Learning, V1, P81, DOI 10.1023/A:1022643204877