Scientific Inference with Interpretable Machine Learning: Analyzing Models to Learn About Real-World Phenomena

被引:2
作者
Freiesleben, Timo [1 ]
Koenig, Gunnar [1 ]
Molnar, Christoph
Tejero-Cantero, Alvaro [1 ]
机构
[1] Univ Tubingen, Cluster Excellence Machine Learning Sci, Maria von Linden Str 6, D-72076 Tubingen, Germany
关键词
Scientific modeling; Interpretable machine learning; Scientific representation; Inference; XAI; IML; NEURAL-NETWORKS; PREDICTION; REPRESENTATION;
D O I
10.1007/s11023-024-09691-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
To learn about real world phenomena, scientists have traditionally used models with clearly interpretable elements. However, modern machine learning (ML) models, while powerful predictors, lack this direct elementwise interpretability (e.g. neural network weights). Interpretable machine learning (IML) offers a solution by analyzing models holistically to derive interpretations. Yet, current IML research is focused on auditing ML models rather than leveraging them for scientific inference. Our work bridges this gap, presenting a framework for designing IML methods-termed 'property descriptors'-that illuminate not just the model, but also the phenomenon it represents. We demonstrate that property descriptors, grounded in statistical learning theory, can effectively reveal relevant properties of the joint probability distribution of the observational data. We identify existing IML methods suited for scientific inference and provide a guide for developing new descriptors with quantified epistemic uncertainty. Our framework empowers scientists to harness ML models for inference, and provides directions for future IML research to support scientific understanding.
引用
收藏
页数:39
相关论文
共 196 条
  • [1] Explaining individual predictions when features are dependent: More accurate approximations to Shapley values
    Aas, Kjersti
    Jullum, Martin
    Loland, Anders
    [J]. ARTIFICIAL INTELLIGENCE, 2021, 298
  • [2] Achinstein P., 1968, Concepts of science: A philosophical analysis, DOI [10.56021/9780801800009, DOI 10.56021/9780801800009]
  • [3] Adebayo J, 2018, ADV NEUR IN, V31
  • [4] Evaluating Saliency Map Explanations for Convolutional Neural Networks: A User Study
    Alqaraawi, Ahmed
    Schuessler, Martin
    Weiss, Philipp
    Costanza, Enrico
    Berthouze, Nadia
    [J]. PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, IUI 2020, 2020, : 275 - 285
  • [5] Anderson C, 2008, WIRED, V16
  • [6] Visualizing the effects of predictor variables in black box supervised learning models
    Apley, Daniel W.
    Zhu, Jingyu
    [J]. JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 2020, 82 (04) : 1059 - 1086
  • [7] Arjovsky M, 2020, Arxiv, DOI [arXiv:1907.02893, 10.48550/arXiv.1907.02893]
  • [8] Bailer-Jones D. M., 2003, INT STUD PHILOS SCI, V17, P59, DOI DOI 10.1080/02698590305238
  • [9] Bailer-Jones DM., 2002, Modeling data: Analogies in neural networks, simulated annealing and genetic algorithms, model-based reasoning, P147
  • [10] Bareinboim E, 2015, ADV NEUR IN, V28