A quantitative evaluation of explainable AI methods using the depth of decision tree

被引:4
作者
Ahmed, Nizar Abdulaziz Mahyoub [1 ]
Alpkocak, Adil [2 ]
机构
[1] Dokuz Eylul Univ, Dept Comp Engn, Izmir, Turkey
[2] Izmir Bakircay Univ, Dept Comp Engn, Izmir, Turkey
关键词
Explainable AI; medical multiclass classification; SHAP; LIME; decision tree; quantitative explainability evaluation;
D O I
10.55730/1300-0632.3924
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is necessary to develop an explainable model to clarify how and why a medical model makes a particular decision. Local posthoc explainable AI (XAI) techniques, such as SHAP and LIME, interpret classification system predictions by displaying the most important features and rules underlying any prediction locally. Therefore, in order to compare two or more XAI methods, they must first be evaluated qualitatively or quantitatively. This paper proposes quantitative XAI evaluation metrics that are not based on biased and subjective human judgment. On the other hand, it is dependent on the depth of the decision tree (DT) to automatically and effectively measure the complexity of XAI methods. Our study introduces a novel XAI strategy that measures the complexity of any XAI method by using a characteristic of another model as a proxy. The output of XAI methods, specifically feature importance scores from SHAP and LIME, is fed into the DT in our proposal. The DT will then draw a full tree based on the feature importance score decisions. As a result, we developed two main metrics that can be used to assess the DT's complexity and thus the associated XAI method: the total depth of the tree (TDT) and the average of the weighted class depth (ACD). The results show that SHAP outperforms LIME and is thus less complex. Furthermore, in terms of the number of documents and features, SHAP is more scalable. These results can indicate whether a specific XAI method is suitable for dealing with different document scales. Furthermore, they can demonstrate which features can be used to improve the performance of the black-box model, in this case, a feedforward neural network (FNN).
引用
收藏
页码:2054 / 2072
页数:20
相关论文
共 35 条
[31]   Explainable Deep Learning Models in Medical Image Analysis [J].
Singh, Amitojdeep ;
Sengupta, Sourya ;
Lakshminarayanan, Vasudevan .
JOURNAL OF IMAGING, 2020, 6 (06)
[32]   Deep Neural Network-Based Screening Model for COVID-19-Infected Patients Using Chest X-Ray Images [J].
Singh, Dilbag ;
Kumar, Vijay ;
Yadav, Vaishali ;
Kaur, Manjit .
INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2021, 35 (03)
[33]   Explainable Prediction of Medical Codes With Knowledge Graphs [J].
Teng, Fei ;
Yang, Wei ;
Chen, Li ;
Huang, LuFei ;
Xu, Qiang .
FRONTIERS IN BIOENGINEERING AND BIOTECHNOLOGY, 2020, 8
[34]   Evaluating XAI: A comparison of rule-based and example-based explanations [J].
van der Waa, Jasper ;
Nieuwburg, Elisabeth ;
Cremers, Anita ;
Neerincx, Mark .
ARTIFICIAL INTELLIGENCE, 2021, 291
[35]   Evaluating the Quality of Machine Learning Explanations: A Survey on Methods and Metrics [J].
Zhou, Jianlong ;
Gandomi, Amir H. ;
Chen, Fang ;
Holzinger, Andreas .
ELECTRONICS, 2021, 10 (05) :1-19