New explainability method for BERT-based model in fake news detection

被引:63
作者
Szczepanski, Mateusz [1 ,2 ]
Pawlicki, Marek [1 ,2 ]
Kozik, Rafal [1 ,2 ]
Choras, Michal [1 ,2 ]
机构
[1] ITTI Sp Zoo, Poznan, Poland
[2] Bydgoszcz Univ Sci & Technol PBS, Bydgoszcz, Poland
基金
欧盟地平线“2020”;
关键词
SCIENCE;
D O I
10.1038/s41598-021-03100-6
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
The ubiquity of social media and their deep integration in the contemporary society has granted new ways to interact, exchange information, form groups, or earn money-all on a scale never seen before. Those possibilities paired with the widespread popularity contribute to the level of impact that social media display. Unfortunately, the benefits brought by them come at a cost. Social Media can be employed by various entities to spread disinformation-so called 'Fake News', either to make a profit or influence the behaviour of the society. To reduce the impact and spread of Fake News, a diverse array of countermeasures were devised. These include linguistic-based approaches, which often utilise Natural Language Processing (NLP) and Deep Learning (DL). However, as the latest advancements in the Artificial Intelligence (AI) domain show, the model's high performance is no longer enough. The explainability of the system's decision is equally crucial in real-life scenarios. Therefore, the objective of this paper is to present a novel explainability approach in BERT-based fake news detectors. This approach does not require extensive changes to the system and can be attached as an extension for operating detectors. For this purposes, two Explainable Artificial Intelligence (xAI) techniques, Local Interpretable Model-Agnostic Explanations (LIME) and Anchors, will be used and evaluated on fake news data, i.e., short pieces of text forming tweets or headlines. This focus of this paper is on the explainability approach for fake news detectors, as the detectors themselves were part of previous works of the authors.
引用
收藏
页数:13
相关论文
共 61 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]   Detecting opinion spams and fake news using text classification [J].
Ahmed, Hadeer ;
Traore, Issa ;
Saad, Sherif .
SECURITY AND PRIVACY, 2018, 1 (01)
[3]   Social Media and Fake News in the 2016 Election [J].
Allcott, Hunt ;
Gentzkow, Matthew .
JOURNAL OF ECONOMIC PERSPECTIVES, 2017, 31 (02) :211-235
[4]   Sentiment Analysis for Fake News Detection [J].
Alonso, Miguel A. ;
Vilares, David ;
Gomez-Rodriguez, Carlos ;
Vilares, Jesus .
ELECTRONICS, 2021, 10 (11)
[5]   Exposure to ideologically diverse news and opinion on Facebook [J].
Bakshy, Eytan ;
Messing, Solomon ;
Adamic, Lada A. .
SCIENCE, 2015, 348 (6239) :1130-1132
[6]   Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI [J].
Barredo Arrieta, Alejandro ;
Diaz-Rodriguez, Natalia ;
Del Ser, Javier ;
Bennetot, Adrien ;
Tabik, Siham ;
Barbado, Alberto ;
Garcia, Salvador ;
Gil-Lopez, Sergio ;
Molina, Daniel ;
Benjamins, Richard ;
Chatila, Raja ;
Herrera, Francisco .
INFORMATION FUSION, 2020, 58 :82-115
[7]  
Bisaillon C, 2020, KAGGLE FAKE REAL NEW
[8]   Machine Learning Explainability Through Comprehensible Decision Trees [J].
Blanco-Justicia, Alberto ;
Domingo-Ferrer, Josep .
MACHINE LEARNING AND KNOWLEDGE EXTRACTION, CD-MAKE 2019, 2019, 11713 :15-26
[9]  
Cantarella M., 2019, DEMB WORKING PAPER S, V146
[10]  
Castelvecchi D, 2016, NATURE, V538, P21, DOI [10.1038/nature.2016.20491, 10.1038/538020a]