OpenXAI: Towards a Transparent Evaluation of Post hoc Model Explanations

被引:0
作者
Agarwal, Chirag [1 ,2 ]
Krishna, Satyapriya [1 ]
Saxena, Eshika [1 ]
Pawelczyk, Martin [3 ]
Johnson, Nari [4 ]
Puri, Isha [1 ]
Zitnik, Marinka [1 ]
Lakkaraju, Himabindu [1 ]
机构
[1] Harvard Univ, Cambridge, MA 02138 USA
[2] Adobe, San Jose, CA 95110 USA
[3] Univ Tubingen, Tubingen, Germany
[4] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
来源
ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022) | 2022年
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While several types of post hoc explanation methods have been proposed in recent literature, there is very little work on systematically benchmarking these methods. Here, we introduce OpenXAI, a comprehensive and extensible open-source framework for evaluating and benchmarking post hoc explanation methods. OpenXAI comprises of the following key components: (i) a flexible synthetic data generator and a collection of diverse real-world datasets, pre-trained models, and state-of-the-art feature attribution methods, (ii) open-source implementations of twenty-two quantitative metrics for evaluating faithfulness, stability (robustness), and fairness of explanation methods, and (iii) the first ever public XAI leaderboards to readily compare several explanation methods across a wide variety of metrics, models, and datasets. OpenXAI is easily extensible, as users can readily evaluate custom explanation methods and incorporate them into our leaderboards. Overall, OpenXAI provides an automated end-to-end pipeline that not only simplifies and standardizes the evaluation of post hoc explanation methods, but also promotes transparency and reproducibility in benchmarking these methods. While the first release of OpenXAI supports only tabular datasets, the explanation methods and metrics that we consider are general enough to be applicable to other data modalities. OpenXAI datasets and data loaders, implementations of state-of-the-art explanation methods and evaluation metrics, as well as leaderboards are publicly available at https://open- xai.github.io/. OpenXAI will be regularly updated to incorporate text and image datasets, other new metrics and explanation methods, and welcomes inputs from the community.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Normalizing trust: Participants' immediately post-hoc explanations of behaviour in Milgram's "obedience' experiments
    Hollander, Matthew M.
    Turowetz, Jason
    BRITISH JOURNAL OF SOCIAL PSYCHOLOGY, 2017, 56 (04) : 655 - 674
  • [42] Towards a transparent data access model for the GridRPC paradigm
    Antoniu, Gabriel
    Caron, Eddy
    Desprez, Frederic
    Fevre, Aurelia
    Jan, Mathieu
    HIGH PERFORMANCE COMPUTING - HIPC 2007, PROCEEDINGS, 2007, 4873 : 269 - +
  • [43] XGNN: Towards Model-Level Explanations of Graph Neural Networks
    Yuan, Hao
    Tang, Jiliang
    Hu, Xia
    Ji, Shuiwang
    KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 430 - 438
  • [44] Towards a Unified Model for Generating Answers and Explanations in Visual Question Answering
    Whitehouse, Chenxi
    Weyde, Tillman
    Madhyastha, Pranava
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1693 - 1705
  • [45] Author Correction: Using model explanations to guide deep learning models towards consistent explanations for EHR data
    Matthew Watson
    Bashar Awwad Shiekh Hasan
    Noura Al Moubayed
    Scientific Reports, 13
  • [46] Towards Realistic Performance Evaluation of Mobile Ad hoc Network
    Kumar, Chirag
    Nagpal, C. K.
    Bhushan, Bharat
    Gupta, Shailender
    PROCEEDINGS OF THE 2012 WORLD CONGRESS ON INFORMATION AND COMMUNICATION TECHNOLOGIES, 2012, : 287 - 291
  • [47] Enhancing Trust in Alzheimer's Disease Classification using Explainable Artificial Intelligence: Incorporating Local Post Hoc Explanations for a Glass-box Model
    Varghese, Abraham
    George, Ben
    Sherimon, Vinu
    Al Shuaily, Huda Salim
    BAHRAIN MEDICAL BULLETIN, 2023, 45 (02) : 1471 - 1478
  • [48] Towards a more transparent and research-based evaluation of design
    van der Voordt, TJM
    RESEARCH BY DESIGN, CONFERENCE PROCEEDINGS B, 2001, : 336 - 343
  • [49] Diffusion-Based Visual Counterfactual Explanations - Towards Systematic Quantitative Evaluation
    Vaeth, Philipp
    Fruehwald, Alexander M.
    Paassen, Benjamin
    Gregorova, Magda
    MACHINE LEARNING AND PRINCIPLES AND PRACTICE OF KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2023, PT III, 2025, 2135 : 120 - 135
  • [50] POST-HOC, ERGO PROPTER HOC - SUBJECTIVE EVALUATION OF EFFECTIVENESS OF 3 TYPES OF THERAPY
    LIEBHART, EH
    ZEITSCHRIFT FUR KLINISCHE PSYCHOLOGIE-FORSCHUNG UND PRAXIS, 1978, 7 (03): : 172 - 185