Intra and Inter-modality Incongruity Modeling and Adversarial Contrastive Learning for Multimodal Fake News Detection

被引:0
作者
Wei, Siqi [1 ]
Wu, Bin [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Beijing, Peoples R China
来源
PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024 | 2024年
基金
中国国家自然科学基金;
关键词
Multimodal Fake News Detection; Intra and Inter-modality Incongruity Modeling; Adversarial Data argumentation; Contrastive Learning;
D O I
10.1145/3652583.3658118
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multimodal fake news detection (FND) is significant in safeguarding network security and societal safety. Most existing studies only focus on common semantic features between different modalities and utilize simple cross-entropy loss for model training. However, these studies overlook the incongruent semantic features in multimodal news data, which can arise within or between modalities. Moreover, the utilization of simple cross-entropy loss may not provide the model with robustness against well-designed forged fake news. To address the above issues, we propose a novel approach named Signed Attention-based Graph Transformer with Adversarial Contrastive Learning (SAGT-ACL) for the detection of multimodal fake news. SAGT-ACL models fine-grained semantic associations in multimodal news articles by constructing a fully connected multimodal graph and reframes the fake news classification task as a graph classification problem. Additionally, SAGT-ACL incorporates a signed attention-based graph transformer module to identify both common and incongruent semantics within and across modalities. Finally, SAGT-ACL proposes an adversarial data augmentation mechanism to simulate malicious forgeries by fake news creators and designs an auxiliary adversarial contrastive learning task to help the model learn more discriminative news representations from the adversarial samples for robust and effective detection. Extensive experiments demonstrate that SAGT-ACL outperforms existing methods, with detection accuracy improvements of 4.95%, 6.01%, and 5.68% on Weibo, Twitter, and Gossipcop datasets, respectively.
引用
收藏
页码:666 / 674
页数:9
相关论文
共 33 条
[1]  
Alam F., 2022, P 29 INT C COMP LING, P6625
[2]   Detection and visualization of misleading content on Twitter [J].
Boididou, Christina ;
Papadopoulos, Symeon ;
Zampoglou, Markos ;
Apostolidis, Lazaros ;
Papadopoulou, Olga ;
Kompatsiaris, Yiannis .
INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2018, 7 (01) :71-86
[3]   Is Multi-Modal Necessarily Better? Robustness Evaluation of Multi-Modal Fake News Detection [J].
Chen, Jinyin ;
Jia, Chengyu ;
Zheng, Haibin ;
Chen, Ruoxi ;
Fu, Chenbo .
IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (06) :3144-3158
[4]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171
[5]  
Dosovitskiy A., 2020, INT C LEARN REPR, P1
[6]   Deep learning for fake news detection: A comprehensive survey [J].
Hu, Linmei ;
Wei, Siqi ;
Zhao, Ziwang ;
Wu, Bin .
AI OPEN, 2022, 3 :133-155
[7]   Multimodal Fusion with Recurrent Neural Networks for Rumor Detection on Microblogs [J].
Jin, Zhiwei ;
Cao, Juan ;
Guo, Han ;
Zhang, Yongdong ;
Luo, Jiebo .
PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, :795-803
[8]  
Kai Shu, 2017, ACM SIGKDD Explorations Newsletter, V19, P22, DOI 10.1145/3137597.3137600
[9]   MVAE: Multimodal Variational Autoencoder for Fake News Detection [J].
Khattar, Dhruv ;
Goud, Jaipal Singh ;
Gupta, Manish ;
Varma, Vasudeva .
WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, :2915-2921
[10]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001