Multi-modal Fake News Detection on Social Media via Multi-grained Information Fusion

被引:0
作者
Zhou, Yangming [1 ]
Yang, Yuzhou [1 ]
Ying, Qichao [1 ]
Qian, Zhenxing [2 ]
Zhang, Xinpeng [2 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai, Peoples R China
[2] Fudan Univ, Sch Comp Sci, Key Lab Culture & Tourism Intelligent Comp, Minist Culture & Tourism, Shanghai, Peoples R China
来源
PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023 | 2023年
基金
中国国家自然科学基金;
关键词
Fake news detection; Multi-modal learning; Multi-modal fusion;
D O I
10.1145/3591106.3592271
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The easy sharing of multimedia content on social media has caused a rapid dissemination of fake news, which threatens society's stability and security. Therefore, fake news detection has garnered extensive research interest in the field of social forensics. Current methods primarily concentrate on the integration of textual and visual features but fail to effectively exploit multi-modal information at both fine-grained and coarse-grained levels. Furthermore, they suffer from an ambiguity problem due to a lack of correlation between modalities or a contradiction between the decisions made by each modality. To overcome these challenges, we present a Multi-grained Multi-modal Fusion Network (MMFN) for fake news detection. Inspired by the multi-grained process of human assessment of news authenticity, we respectively employ two Transformer-based pre-trained models to encode token-level features from text and images. The multi-modal module fuses fine-grained features, taking into account coarse-grained features encoded by the CLIP encoder. To address the ambiguity problem, we design uni-modal branches with similarity-based weighting to adaptively adjust the use of multi-modal features. Experimental results demonstrate that the proposed framework outperforms state-of-the-art methods on three prevalent datasets.
引用
收藏
页码:343 / 352
页数:10
相关论文
共 48 条
[1]   Open-Domain, Content-based, Multi-modal Fact-checking of Out-of-Context Images via Online Resources [J].
Abdelnabi, Sahar ;
Hasan, Rakibul ;
Fritz, Mario .
2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, :14920-14929
[2]  
Allein L, 2021, arXiv
[3]   Multimodal Machine Learning: A Survey and Taxonomy [J].
Baltrusaitis, Tadas ;
Ahuja, Chaitanya ;
Morency, Louis-Philippe .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) :423-443
[4]   Combining Neural, Statistical and External Features for Fake News Stance Identification [J].
Bhatt, Gaurav ;
Sharma, Aman ;
Sharma, Shivam ;
Nagpal, Ankush ;
Raman, Balasubramanian ;
Mittal, Ankush .
COMPANION PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE 2018 (WWW 2018), 2018, :1353-1357
[5]  
Bian T, 2020, AAAI CONF ARTIF INTE, V34, P549
[6]   Detection and visualization of misleading content on Twitter [J].
Boididou, Christina ;
Papadopoulos, Symeon ;
Zampoglou, Markos ;
Apostolidis, Lazaros ;
Papadopoulou, Olga ;
Kompatsiaris, Yiannis .
INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2018, 7 (01) :71-86
[7]   Cross-modal Ambiguity Learning for Multimodal Fake News Detection [J].
Chen, Yixuan ;
Li, Dongsheng ;
Zhang, Peng ;
Sui, Jie ;
Lv, Qin ;
Lu, Tun ;
Shang, Li .
PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, :2897-2905
[8]   CLIP-Art: Contrastive Pre-training for Fine-Grained Art Classification [J].
Conde, Marcos, V ;
Turgutlu, Kerem .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, :3951-3955
[9]  
Conroy NK., 2015, Proceedings of the 78th ASIS T Annual Meeting: Information Science with Impact: Research in and for the Community, V82, P1, DOI [10.1002/pra2.2015.145052010082, DOI 10.1002/PRA2.2015.145052010082]
[10]   Beyond Question-Based Biases: Assessing Multimodal Shortcut Learning in Visual Question Answering [J].
Dancette, Corentin ;
Cadene, Remi ;
Teney, Damien ;
Cord, Matthieu .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :1554-1563