Toward interpretable machine learning: evaluating models of heterogeneous predictions

被引:2
|
作者
Zhang, Ruixun [1 ,2 ,3 ,4 ]
机构
[1] Peking Univ, Sch Math Sci, Beijing, Peoples R China
[2] Peking Univ, Ctr Stat Sci, Beijing, Peoples R China
[3] Peking Univ, Natl Engn Lab Big Data Anal & Applicat, Beijing, Peoples R China
[4] Peking Univ, Lab Math Econ & Quantitat Finance, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Machine learning; Interpretability; Heterogeneous prediction; Bayesian statistics; Loan default; SYSTEMIC RISK; FINANCE; DEFAULT; GAME; GO;
D O I
10.1007/s10479-024-06033-1
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
AI and machine learning have made significant progress in the past decade, powering many applications in FinTech and beyond. But few machine learning models, especially deep learning models, are interpretable by humans, creating challenges for risk management and model improvements. Here, we propose a simple yet powerful framework to evaluate and interpret any black-box model with binary outcomes and explanatory variables, and heterogeneous relationships between the two. Our new metric, the signal success share (SSS) cross-entropy loss, measures how well the model captures the relationship along any feature or dimension, thereby providing actionable guidance on model improvements. Simulations demonstrate that our metric works for heterogeneous and nonlinear predictions, and distinguishes itself from traditional loss functions in evaluating model interpretability. We apply the methodology to an example of predicting loan defaults with real data. Our framework is more broadly applicable to a wide range of problems in financial and information technology.
引用
收藏
页码:867 / 887
页数:21
相关论文
共 50 条
  • [21] Interpretable machine learning models for concrete compressive strength prediction
    Hoang, Huong-Giang Thi
    Nguyen, Thuy-Anh
    Ly, Hai-Bang
    INNOVATIVE INFRASTRUCTURE SOLUTIONS, 2025, 10 (01)
  • [22] Interpretable machine learning models for failure cause prediction in imbalanced oil pipeline data
    Awuku, Bright
    Huang, Ying
    Yodo, Nita
    Asa, Eric
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2024, 35 (07)
  • [23] A Survey of Machine Learning Models in Renewable Energy Predictions
    Lai, Jung-Pin
    Chang, Yu-Ming
    Chen, Chieh-Huang
    Pai, Ping-Feng
    APPLIED SCIENCES-BASEL, 2020, 10 (17):
  • [24] Application of interpretable machine learning models to improve the prediction performance of ionic liquids toxicity
    Fan, Dingchao
    Xue, Ke
    Zhang, Runqi
    Zhu, Wenguang
    Zhang, Hongru
    Qi, Jianguang
    Zhu, Zhaoyou
    Wang, Yinglong
    Cui, Peizhe
    SCIENCE OF THE TOTAL ENVIRONMENT, 2024, 908
  • [25] Evaluating Machine Learning Models for QoT Estimation
    Morais, Rui Manuel
    Pedro, Joao
    2018 20TH ANNIVERSARY INTERNATIONAL CONFERENCE ON TRANSPARENT OPTICAL NETWORKS (ICTON), 2018,
  • [26] Interpretable Machine Learning with Boosting by Boolean Algorithm
    Neuhaus, Nathan
    Kovalerchuk, Boris
    2019 JOINT 8TH INTERNATIONAL CONFERENCE ON INFORMATICS, ELECTRONICS & VISION (ICIEV) AND 2019 3RD INTERNATIONAL CONFERENCE ON IMAGING, VISION & PATTERN RECOGNITION (ICIVPR) WITH INTERNATIONAL CONFERENCE ON ACTIVITY AND BEHAVIOR COMPUTING (ABC), 2019, : 307 - 311
  • [27] Definitions, methods, and applications in interpretable machine learning
    Murdoch, W. James
    Singh, Chandan
    Kumbier, Karl
    Abbasi-Asl, Reza
    Yu, Bin
    PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (44) : 22071 - 22080
  • [28] Interpretable machine learning to forecast hypoxia in a lagoon
    Politikos, Dimitris, V
    Petasis, Georgios
    Katselis, George
    ECOLOGICAL INFORMATICS, 2021, 66
  • [29] Interpretable machine learning for dementia: A systematic review
    Martin, Sophie A.
    Townend, Florence J.
    Barkhof, Frederik
    Cole, James H.
    ALZHEIMERS & DEMENTIA, 2023, 19 (05) : 2135 - 2149
  • [30] Evaluating Explainable Machine Learning Models for Clinicians
    Scarpato, Noemi
    Nourbakhsh, Aria
    Ferroni, Patrizia
    Riondino, Silvia
    Roselli, Mario
    Fallucchi, Francesca
    Barbanti, Piero
    Guadagni, Fiorella
    Zanzotto, Fabio Massimo
    COGNITIVE COMPUTATION, 2024, 16 (04) : 1436 - 1446