OpenXAI: Towards a Transparent Evaluation of Post hoc Model Explanations

被引:0
|
作者
Agarwal, Chirag [1 ,2 ]
Krishna, Satyapriya [1 ]
Saxena, Eshika [1 ]
Pawelczyk, Martin [3 ]
Johnson, Nari [4 ]
Puri, Isha [1 ]
Zitnik, Marinka [1 ]
Lakkaraju, Himabindu [1 ]
机构
[1] Harvard Univ, Cambridge, MA 02138 USA
[2] Adobe, San Jose, CA 95110 USA
[3] Univ Tubingen, Tubingen, Germany
[4] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While several types of post hoc explanation methods have been proposed in recent literature, there is very little work on systematically benchmarking these methods. Here, we introduce OpenXAI, a comprehensive and extensible open-source framework for evaluating and benchmarking post hoc explanation methods. OpenXAI comprises of the following key components: (i) a flexible synthetic data generator and a collection of diverse real-world datasets, pre-trained models, and state-of-the-art feature attribution methods, (ii) open-source implementations of twenty-two quantitative metrics for evaluating faithfulness, stability (robustness), and fairness of explanation methods, and (iii) the first ever public XAI leaderboards to readily compare several explanation methods across a wide variety of metrics, models, and datasets. OpenXAI is easily extensible, as users can readily evaluate custom explanation methods and incorporate them into our leaderboards. Overall, OpenXAI provides an automated end-to-end pipeline that not only simplifies and standardizes the evaluation of post hoc explanation methods, but also promotes transparency and reproducibility in benchmarking these methods. While the first release of OpenXAI supports only tabular datasets, the explanation methods and metrics that we consider are general enough to be applicable to other data modalities. OpenXAI datasets and data loaders, implementations of state-of-the-art explanation methods and evaluation metrics, as well as leaderboards are publicly available at https://open- xai.github.io/. OpenXAI will be regularly updated to incorporate text and image datasets, other new metrics and explanation methods, and welcomes inputs from the community.
引用
收藏
页数:16
相关论文
共 50 条
  • [11] How can I choose an explainer? An Application-grounded Evaluation of Post-hoc Explanations
    Jesus, Sergio
    Belem, Catarina
    Balayan, Vladimir
    Bento, Joao
    Saleiro, Pedro
    Bizarro, Pedro
    Gama, Joao
    PROCEEDINGS OF THE 2021 ACM CONFERENCE ON FAIRNESS, ACCOUNTABILITY, AND TRANSPARENCY, FACCT 2021, 2021, : 805 - 815
  • [12] Integrating ERP using EAI:: a model for post hoc evaluation
    Sharif, AM
    Irani, Z
    Love, PED
    EUROPEAN JOURNAL OF INFORMATION SYSTEMS, 2005, 14 (02) : 162 - 174
  • [13] Comparing Strategies for Post-Hoc Explanations in Machine Learning Models
    Vij, Aabhas
    Nanjundan, Preethi
    MOBILE COMPUTING AND SUSTAINABLE INFORMATICS, 2022, 68 : 585 - 592
  • [14] Evaluating Stability of Post-hoc Explanations for Business Process Predictions
    Velmurugan, Mythreyi
    Ouyang, Chun
    Moreira, Catarina
    Sindhgatta, Renuka
    SERVICE-ORIENTED COMPUTING (ICSOC 2021), 2021, 13121 : 49 - 64
  • [15] Post Hoc Explanations of Language Models Can Improve Language Models
    Krishna, Satyapriya
    Ma, Jiaqi
    Slack, Dylan
    Ghandeharioun, Asma
    Singh, Sameer
    Lakkaraju, Himabindu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [16] A Study on Trust in Black Box Models and Post-hoc Explanations
    El Bekri, Nadia
    Kling, Jasmin
    Huber, Marco F.
    14TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING MODELS IN INDUSTRIAL AND ENVIRONMENTAL APPLICATIONS (SOCO 2019), 2020, 950 : 35 - 46
  • [17] Post-hoc vs ante-hoc explanations: xAI design guidelines for data scientists
    Retzlaff, Carl O.
    Angerschmid, Alessa
    Saranti, Anna
    Schneeberger, David
    Roettger, Richard
    Mueller, Heimo
    Holzinger, Andreas
    COGNITIVE SYSTEMS RESEARCH, 2024, 86
  • [18] Preference-based and local post-hoc explanations for recommender systems
    Brunot, Leo
    Canovas, Nicolas
    Chanson, Alexandre
    Labroche, Nicolas
    Verdeaux, Willeme
    INFORMATION SYSTEMS, 2022, 108
  • [19] Fairness via Explanation Quality: Evaluating Disparities in the Quality of Post hoc Explanations
    Dai, Jessica
    Upadhyay, Sohini
    Aivodji, Ulrich
    Bach, Stephen H.
    Lakkaraju, Himabindu
    PROCEEDINGS OF THE 2022 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, AIES 2022, 2022, : 203 - 214
  • [20] A general framework for personalising post hoc explanations through user knowledge integration
    Jeyasothy, Adulam
    Laugel, Thibault
    Lesot, Marie-Jeanne
    Marsala, Christophe
    Detyniecki, Marcin
    INTERNATIONAL JOURNAL OF APPROXIMATE REASONING, 2023, 160