Few-shot learning with distribution calibration for event-level rumor detection

被引:0
作者
Ran, Hongyan [1 ]
Jia, Caiyan [2 ,3 ]
Li, Xiaohong [1 ]
Zhang, Zhichang [1 ]
机构
[1] Northwest Normal Univ, Coll Comp Sci & Engn, Lanzhou 730070, Peoples R China
[2] Beijing Jiaotong Univ, Sch Comp & Informat Technol, Beijing 100044, Peoples R China
[3] Beijing Jiaotong Univ, Beijing Key Lab Traff Data Anal & Min, Beijing 100044, Peoples R China
关键词
Rumor detection; Newly emergent events; Few-shot setting; Event-invariant; Calibrate distribution; NETWORK;
D O I
10.1016/j.neucom.2024.129034
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the rapid evolution of social media, rumors travel at unprecedented speeds. Automatic recognition of rumors is important for making users receive truthful information and maintaining social harmony. Recently, deep learning models have demonstrated strong rumor detection ability by capturing semantics of rumors, learning rumor propagation patterns and integrating users' properties, etc. However, most existing rumor detection models perform poorly on unseen events because they are powerful at capturing event-specific features in seen data, which cannot be transferred to newly emergent events due to distribution differences between events. Therefore, in this study, we propose a novel model named E-Rumor in a few-shot setting to learn and transfer event-invariant features from historic events with sufficient samples to new events with only a few examples. The model first calculates the rumor class distributions of historic rumors, and then calibrates the class distributions of anew event with the old ones. Furthermore, an adequate number of samples are generated from the calibrated distributions to expand the training set for a new-event rumor classifier. The proposed model can be paired with any rumor detection classifier as a feature extractor without extra parameters. Empirical studies have shown that a simple MLP (Multilayer Perceptron) trained on the samples generated from the calibrated distributions can outperform the state-of-the-art baseline models on two benchmark rumor datasets PHEME5, PHEME9 and one topic-level dataset T-Twitter, which is generated by topic extraction and cross-data manner, with more than 15.15%, 4.7% and 7.38% accuracy improvements, respectively.
引用
收藏
页数:12
相关论文
共 49 条
[1]  
Antoniou A, 2019, Arxiv, DOI arXiv:1902.09884
[2]  
Bian T, 2020, AAAI CONF ARTIF INTE, V34, P549
[3]   Latent Dirichlet allocation [J].
Blei, DM ;
Ng, AY ;
Jordan, MI .
JOURNAL OF MACHINE LEARNING RESEARCH, 2003, 3 (4-5) :993-1022
[4]  
Castillo C., 2011, Proceedings of the 20th international conference on World wide web, WWW'11, (New York, NY, USA), P675, DOI [DOI 10.1145/1963405.1963500, https://doi.org/10.1145/1963405.1963500]
[5]  
Feng Song., 2012, Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics: Short Papers, V2, P171
[6]  
Finn C, 2017, PR MACH LEARN RES, V70
[7]   Low-shot Visual Recognition by Shrinking and Hallucinating Features [J].
Hariharan, Bharath ;
Girshick, Ross .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :3037-3046
[8]   Heterogeneous Graph Attention Networks for Early Detection of Rumors on Twitter [J].
Huang, Qi ;
Yu, Junshuai ;
Wu, Jia ;
Wang, Bin .
2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
[9]   Multimodal Fusion with Recurrent Neural Networks for Rumor Detection on Microblogs [J].
Jin, Zhiwei ;
Cao, Juan ;
Guo, Han ;
Zhang, Yongdong ;
Luo, Jiebo .
PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, :795-803
[10]  
Kai Shu, 2017, ACM SIGKDD Explorations Newsletter, V19, P22, DOI 10.1145/3137597.3137600