More transparent and explainable machine learning algorithms are required to provide enhanced and sustainable dataset understanding

被引:1
作者
Wood, David A. [1 ]
机构
[1] DWA Energy Ltd, Lincoln, England
关键词
Dataset interrogation; Optimized data matching; Prediction explainability; Forensic dataset interpretability; Transparent open box (TOB) algorithms; !text type='Python']Python[!/text] coded TOB;
D O I
10.1016/j.ecolmodel.2024.110898
中图分类号
Q14 [生态学(生物生态学)];
学科分类号
071012 ; 0713 ;
摘要
For detailed dataset interrogation and auditing purposes the lack of dataset explainability/transparency of the majority of available machine-learning (ML) models poses limitations. There is a tendency for ML models to focus on prediction speed and accuracy at the expense of transparently revealing dataset relationships. A case is made here to broaden that focus and for ML models to offer alternative configurations tailored to provide more explanations about how individual predictions are derived. Indeed, those striving to achieve sustainable objectives should not rely on opaque ML models and seek transparency as a fundamental objective of good modelling practice (GMP). Doing so tends to boost trust and confidence in the outputs of models relating to complex socioenvironmental systems (SES), particularly those being used to potentially justify controversial social, political and ethical decisions. Currently, the transparent open box algorithms (TOB) are the only ML algorithms available that are configured specifically to routinely provide detailed data record relationships for each of their predictions. This study describes the data mining benefits of the Python-coded optimized data-matching TOB algorithms generally, and when applied to environmental datasets characterized by complex non-linear relationships involving many variables.
引用
收藏
页数:7
相关论文
共 29 条
  • [1] Abdollahi B, 2018, HUM-COMPUT INT-SPRIN, P21, DOI 10.1007/978-3-319-90403-0_2
  • [2] Opening the Black Box: Interpretable Machine Learning for Geneticists
    Azodi, Christina B.
    Tang, Jiliang
    Shiu, Shin-Han
    [J]. TRENDS IN GENETICS, 2020, 36 (06) : 442 - 455
  • [3] Analysis of Explainers of Black Box Deep Neural Networks for Computer Vision: A Survey
    Buhrmester, Vanessa
    Muench, David
    Arens, Michael
    [J]. MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2021, 3 (04): : 966 - 989
  • [4] How the machine 'thinks': Understanding opacity in machine learning algorithms
    Burrell, Jenna
    [J]. BIG DATA & SOCIETY, 2016, 3 (01): : 1 - 12
  • [5] Prediction and analysis of net ecosystem carbon exchange based on gradient boosting regression and random forest
    Cai, Jianchao
    Xu, Kai
    Zhu, Yanhui
    Hu, Fang
    Li, Liuhuan
    [J]. APPLIED ENERGY, 2020, 262
  • [6] Extrapolation and AI transparency: Why machine learning models should reveal when they make decisions beyond their training
    Cao, Xuenan
    Yousefzadeh, Roozbeh
    [J]. BIG DATA & SOCIETY, 2023, 10 (01)
  • [7] Making machine learning trustworthy
    Eshete, Birhanu
    [J]. SCIENCE, 2021, 373 (6556) : 743 - 744
  • [8] Seasonality of ecosystem respiration and gross primary production as derived from FLUXNET measurements
    Falge, E
    Baldocchi, D
    Tenhunen, J
    Aubinet, M
    Bakwin, P
    Berbigier, P
    Bernhofer, C
    Burba, G
    Clement, R
    Davis, KJ
    Elbers, JA
    Goldstein, AH
    Grelle, A
    Granier, A
    Guomundsson, J
    Hollinger, D
    Kowalski, AS
    Katul, G
    Law, BE
    Malhi, Y
    Meyers, T
    Monson, RK
    Munger, JW
    Oechel, W
    Paw, KT
    Pilegaard, K
    Rannik, Ü
    Rebmann, C
    Suyker, A
    Valentini, R
    Wilson, K
    Wofsy, S
    [J]. AGRICULTURAL AND FOREST METEOROLOGY, 2002, 113 (1-4) : 53 - 74
  • [9] Towards Transparency by Design for Artificial Intelligence
    Felzmann, Heike
    Fosch-Villaronga, Eduard
    Lutz, Christoph
    Tamo-Larrieux, Aurelia
    [J]. SCIENCE AND ENGINEERING ETHICS, 2020, 26 (06) : 3333 - 3361
  • [10] Harrell FE, 2015, SPRINGER SER STAT, P311, DOI 10.1007/978-3-319-19425-7_13