Shapley-based feature augmentation

被引:13
作者
Antwarg, Liat [1 ]
Galed, Chen [1 ]
Shimoni, Nathaniel [1 ]
Rokach, Lior [1 ]
Shapira, Bracha [1 ]
机构
[1] Ben Gurion Univ Negev, Dept Software & Informat Syst Engn, David Ben Gurion Blvd 1, IL-8410501 Beer Sheva, Israel
关键词
SHAP; XAI; Shapley values; Feature augmentation; CANCER;
D O I
10.1016/j.inffus.2023.03.010
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Improving the predictive performance of machine learning models is the desired goal in many tasks and domains. The predictive performance of the learning algorithm is directly affected by the input features it receives. Feature augmentation is aimed at enhancing the quality of models by adding informative features to the original data. Explainable AI methods are typically used to explain the results of machine learning models. Recently, these methods have also been used to improve models' predictive performance. In this study, we examine the benefit of incorporating the explanations obtained by an explainable AI method as augmented features. In particular, we propose SFA - Shapley-Based feature augmentation, a two-stage ensemble learning method that uses out-of-fold predictions and their corresponding Shapley values as augmented features for each instance. Shapley values, which are obtained without domain expertise, reflect the importance of the original features to each prediction and consider their interactions with all other features. Experimental results demonstrate the superiority of our proposed method, SFA, against several feature augmentation methods on multiple public datasets with various characteristics.
引用
收藏
页码:92 / 102
页数:11
相关论文
共 55 条
[1]   Explaining individual predictions when features are dependent: More accurate approximations to Shapley values [J].
Aas, Kjersti ;
Jullum, Martin ;
Loland, Anders .
ARTIFICIAL INTELLIGENCE, 2021, 298
[2]  
Adebayo J, 2020, Arxiv, DOI arXiv:2011.05429
[3]   Optuna: A Next-generation Hyperparameter Optimization Framework [J].
Akiba, Takuya ;
Sano, Shotaro ;
Yanase, Toshihiko ;
Ohta, Takeru ;
Koyama, Masanori .
KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, :2623-2631
[4]   On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation [J].
Bach, Sebastian ;
Binder, Alexander ;
Montavon, Gregoire ;
Klauschen, Frederick ;
Mueller, Klaus-Robert ;
Samek, Wojciech .
PLOS ONE, 2015, 10 (07)
[5]  
Bento M., 2021, Discov. Artif. Intell., V1, P1, DOI DOI 10.1007/S44163-021-00008-Y
[6]  
Bobek Szymon, 2021, Computational Science - ICCS 2021. 21st International Conference. Proceedings. Lecture Notes in Computer Science (LNCS 12747), P361, DOI 10.1007/978-3-030-77980-1_28
[7]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[8]   XGBoost: A Scalable Tree Boosting System [J].
Chen, Tianqi ;
Guestrin, Carlos .
KDD'16: PROCEEDINGS OF THE 22ND ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2016, :785-794
[9]   Multi-Level Semantic Feature Augmentation for One-Shot Learning [J].
Chen, Zitian ;
Fu, Yanwei ;
Zhang, Yinda ;
Jiang, Yu-Gang ;
Xue, Xiangyang ;
Sigal, Leonid .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (09) :4594-4605
[10]  
Covert I., 2021, J. Mach. Learn. Res., V22, p209:1