CHIRPS: Explaining random forest classification

被引:0
作者
Julian Hatwell
Mohamed Medhat Gaber
R. Muhammad Atif Azad
机构
[1] Birmingham City University,
来源
Artificial Intelligence Review | 2020年 / 53卷
关键词
XAI; Model interpretability; Random forests; Classification; Frequent patterns;
D O I
暂无
中图分类号
学科分类号
摘要
Modern machine learning methods typically produce “black box” models that are opaque to interpretation. Yet, their demand has been increasing in the Human-in-the-Loop processes, that is, those processes that require a human agent to verify, approve or reason about the automated decisions before they can be applied. To facilitate this interpretation, we propose Collection of High Importance Random Path Snippets (CHIRPS); a novel algorithm for explaining random forest classification per data instance. CHIRPS extracts a decision path from each tree in the forest that contributes to the majority classification, and then uses frequent pattern mining to identify the most commonly occurring split conditions. Then a simple, conjunctive form rule is constructed where the antecedent terms are derived from the attributes that had the most influence on the classification. This rule is returned alongside estimates of the rule’s precision and coverage on the training data along with counter-factual details. An experimental study involving nine data sets shows that classification rules returned by CHIRPS have a precision at least as high as the state of the art when evaluated on unseen data (0.91–0.99) and offer a much greater coverage (0.04–0.54). Furthermore, CHIRPS uniquely controls against under- and over-fitting solutions by maximising novel objective functions that are better suited to the local (per instance) explanation setting.
引用
收藏
页码:5747 / 5788
页数:41
相关论文
共 22 条
[1]  
Andrews R(1995)Survey and critique of techniques for extracting rules from trained artificial neural networks Knowl-Based Syst 8 373-389
[2]  
Diederich J(2012)Analysis of a random forests model J Mach Learn Res 13 1063-1095
[3]  
Tickle AB(2001)Random forests Mach Learn 45 5-32
[4]  
Biau G(2014)Interpreting tree ensembles with intrees Int J Data Sci Anal 7 277-87
[5]  
Breiman L(2014)Do we need hundreds of classifiers to solve real world classification problems J Mach Learn Res 15 3133-3181
[6]  
Deng H(2014)Comprehensible classification models: a position paper ACM SIGKDD Explor Newsl 15 1-10
[7]  
Fernandez-Delgado M(2001)Greedy function approximation: a gradient boosting machine Ann Stat 29 1189-1232
[8]  
Freitas AA(2006)Classifier technology and the illusion of progress Stat Sci 21 1-14
[9]  
Friedman J(1948)Studies in the logic of explanation Philos Sci 15 135-175
[10]  
Hand DJ(2012)The dawn of a critical transparency right for the profiling era Digital Enlight Yearb 2012 41-56