Explanation of machine learning models using shapley additive explanation and application for real data in hospital

被引:264
作者
Nohara, Yasunobu [1 ]
Matsumoto, Koutarou [2 ]
Soejima, Hidehisa [3 ]
Nakashima, Naoki [4 ]
机构
[1] Kumamoto Univ, Kumamoto, Japan
[2] Kurume Univ, Fukuoka, Japan
[3] Saiseikai Kumamoto Hosp, Kumamoto, Japan
[4] Kyushu Univ Hosp, Fukuoka, Japan
关键词
Shapley additive explanation; Machine learning; Interpretability; Feature importance; Feature packing;
D O I
10.1016/j.cmpb.2021.106584
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Background and Objective: When using machine learning techniques in decision-making processes, the interpretability of the models is important. In the present paper, we adopted the Shapley additive explanation (SHAP), which is based on fair profit allocation among many stakeholders depending on their contribution, for interpreting a gradient-boosting decision tree model using hospital data. Methods: For better interpretability, we propose two novel techniques as follows: (1) a new metric of feature importance using SHAP and (2) a technique termed feature packing, which packs multiple similar features into one grouped feature to allow an easier understanding of the model without reconstruction of the model. We then compared the explanation results between the SHAP framework and existing methods using cerebral infarction data from our hospital. Results: The interpretation by SHAP was mostly consistent with that by the existing methods. We showed how the A/G ratio works as an important prognostic factor for cerebral infarction using proposed techniques. Conclusion: Our techniques are useful for interpreting machine learning models and can uncover the underlying relationships between features and outcome. (C) 2021 Elsevier B.V. All rights reserved.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Prognosis Prediction of Stroke based on Machine Learning and Explanation Model
    Qin, Qiuli
    Zhou, Xuehan
    Jiang, Yong
    INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2021, 16 (02) : 1 - 13
  • [32] Combining Machine Learning and Bayesian Networks for ECG Interpretation and Explanation
    Wanyana, Tezira
    Nzomo, Mbithe
    Price, C. Sue
    Moodley, Deshendran
    ICT4AWE: PROCEEDINGS OF THE 8TH INTERNATIONAL CONFERENCE ON INFORMATION AND COMMUNICATION TECHNOLOGIES FOR AGEING WELL AND E-HEALTH, 2022, : 81 - 92
  • [33] Random forest and Shapley additive explanation for compressive strength prediction of NaOH-pretreated crumb rubber concrete
    Sun, Yang
    PROGRESS IN RUBBER PLASTICS AND RECYCLING TECHNOLOGY, 2024, 40 (02) : 151 - 167
  • [34] A Responsible Machine Learning Workflow with Focus on Interpretable Models, Post-hoc Explanation, and Discrimination Testing
    Gill, Navdeep
    Hall, Patrick
    Montgomery, Kim
    Schmidt, Nicholas
    INFORMATION, 2020, 11 (03)
  • [35] Machine learning modeling and additive explanation techniques for glutathione production from multiple experimental growth conditions of Saccharomyces cerevisiae
    Fuhr, Ana Carolina Ferreira Piazzi
    Goncalves, Ingrid da Mata
    Santos, Lucielen Oliveira
    Salau, Nina Paula Goncalves
    INTERNATIONAL JOURNAL OF BIOLOGICAL MACROMOLECULES, 2024, 262
  • [36] Using Machine Learning Models and Actual Transaction Data for Predicting Real Estate Prices
    Pai, Ping-Feng
    Wang, Wen-Chang
    APPLIED SCIENCES-BASEL, 2020, 10 (17):
  • [37] Prediction of HHV of fuel by Machine learning Algorithm: Interpretability analysis using Shapley Additive Explanations (SHAP)
    Timilsina, Manish Sharma
    Sen, Subhadip
    Uprety, Bibek
    Patel, Vashishtha B.
    Sharma, Prateek
    Sheth, Pratik N.
    FUEL, 2024, 357
  • [38] Prediction of HHV of fuel by Machine learning Algorithm: Interpretability analysis using Shapley Additive Explanations (SHAP)
    Timilsina, Manish Sharma
    Sen, Subhadip
    Uprety, Bibek
    Patel, Vashishtha B.
    Sharma, Prateek
    Sheth, Pratik N.
    FUEL, 2024, 357
  • [39] Shifts in Climatic Limitations on Global Vegetation Productivity Unveiled by Shapley Additive Explanation: Reduced Temperature but Increased Water Limitations
    Xie, Jiangliu
    Yin, Gaofei
    Xie, Qiaoyun
    Wu, Chaoyang
    Yuan, Wenping
    Zeng, Yelu
    Verger, Aleixandre
    Descals, Adria
    Filella, Iolanda
    Penuelas, Josep
    JOURNAL OF GEOPHYSICAL RESEARCH-BIOGEOSCIENCES, 2024, 129 (12)
  • [40] Shapley-Additive-Explanations-Based Factor Analysis for Dengue Severity Prediction using Machine Learning
    Chowdhury, Shihab Uddin
    Sayeed, Sanjana
    Rashid, Iktisad
    Alam, Md Golam Rabiul
    Masum, Abdul Kadar Muhammad
    Dewan, M. Ali Akber
    JOURNAL OF IMAGING, 2022, 8 (09)