Automated Machine Learning and Explainable AI (AutoML-XAI) for Metabolomics: Improving Cancer Diagnostics

被引:4
|
作者
Bifarin, Olatomiwa O. [1 ]
Fernandez, Facundo M. [1 ,2 ]
机构
[1] Georgia Inst Technol, Sch Chem & Biochem, Atlanta, GA 30332 USA
[2] Georgia Inst Technol, Petit Inst Bioengn & Biosci, Atlanta, GA 30332 USA
关键词
metabolomics; automated machine learning; explainableAI; cancer biology; Shapley additive explanations;
D O I
10.1021/jasms.3c00403
中图分类号
Q5 [生物化学];
学科分类号
071010 ; 081704 ;
摘要
Metabolomics generates complex data necessitating advanced computational methods for generating biological insight. While machine learning (ML) is promising, the challenges of selecting the best algorithms and tuning hyperparameters, particularly for nonexperts, remain. Automated machine learning (AutoML) can streamline this process; however, the issue of interpretability could persist. This research introduces a unified pipeline that combines AutoML with explainable AI (XAI) techniques to optimize metabolomics analysis. We tested our approach on two data sets: renal cell carcinoma (RCC) urine metabolomics and ovarian cancer (OC) serum metabolomics. AutoML, using Auto-sklearn, surpassed standalone ML algorithms like SVM and k-Nearest Neighbors in differentiating between RCC and healthy controls, as well as OC patients and those with other gynecological cancers. The effectiveness of Auto-sklearn is highlighted by its AUC scores of 0.97 for RCC and 0.85 for OC, obtained from the unseen test sets. Importantly, on most of the metrics considered, Auto-sklearn demonstrated a better classification performance, leveraging a mix of algorithms and ensemble techniques. Shapley Additive Explanations (SHAP) provided a global ranking of feature importance, identifying dibutylamine and ganglioside GM(d34:1) as the top discriminative metabolites for RCC and OC, respectively. Waterfall plots offered local explanations by illustrating the influence of each metabolite on individual predictions. Dependence plots spotlighted metabolite interactions, such as the connection between hippuric acid and one of its derivatives in RCC, and between GM3(d34:1) and GM3(18:1_16:0) in OC, hinting at potential mechanistic relationships. Through decision plots, a detailed error analysis was conducted, contrasting feature importance for correctly versus incorrectly classified samples. In essence, our pipeline emphasizes the importance of harmonizing AutoML and XAI, facilitating both simplified ML application and improved interpretability in metabolomics data science.
引用
收藏
页码:1089 / 1100
页数:12
相关论文
共 50 条
  • [1] Explainable AI (XAI) Applied in Machine Learning for Pain Modeling: A Review
    Madanu, Ravichandra
    Abbod, Maysam F.
    Hsiao, Fu-Jung
    Chen, Wei-Ta
    Shieh, Jiann-Shing
    TECHNOLOGIES, 2022, 10 (03)
  • [2] From Explainable AI to Explainable Simulation: Using Machine Learning and XAI to understand System Robustness
    Feldkamp, Niclas
    Strassburger, Steffen
    PROCEEDINGS OF THE 2023 ACM SIGSIM INTERNATIONAL CONFERENCE ON PRINCIPLES OF ADVANCED DISCRETE SIMULATION, ACMSIGSIM-PADS 2023, 2023, : 96 - 106
  • [3] Unmasking Banking Fraud: Unleashing the Power of Machine Learning and Explainable AI (XAI) on Imbalanced Data
    Nobel, S. M. Nuruzzaman
    Sultana, Shirin
    Singha, Sondip Poul
    Chaki, Sudipto
    Mahi, Md. Julkar Nayeen
    Jan, Tony
    Barros, Alistair
    Whaiduzzaman, Md
    INFORMATION, 2024, 15 (06)
  • [4] Using an Ensemble Machine Learning Model with Explainable AI (XAI) to Diagnose Gestational Diabetes Mellitus
    Pasha, S. T.
    Sikder, S.
    Rahman, M. M.
    Islam, A.
    Alam, M. Z.
    Habib, M. T.
    Amin, M. A.
    DIABETES RESEARCH AND CLINICAL PRACTICE, 2024, 209
  • [5] Diagnostics of Thyroid Cancer Using Machine Learning and Metabolomics
    Kuang, Alyssa
    Kouznetsova, Valentina L.
    Kesari, Santosh
    Tsigelny, Igor F.
    METABOLITES, 2024, 14 (01)
  • [6] Can surgeons trust AI? Perspectives on machine learning in surgery and the importance of eXplainable Artificial Intelligence (XAI)
    Brandenburg, Johanna M.
    Mueller-Stich, Beat P.
    Wagner, Martin
    van der Schaar, Mihaela
    LANGENBECKS ARCHIVES OF SURGERY, 2025, 410 (01)
  • [7] A Review on Automated Machine Learning (AutoML) Systems
    Nagarajah, Thiloshon
    Poravi, Guhanathan
    2019 IEEE 5TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2019,
  • [8] Explainable Machine Learning for Trustworthy AI
    Giannotti, Fosca
    ARTIFICIAL INTELLIGENCE RESEARCH AND DEVELOPMENT, 2022, 356 : 3 - 3
  • [9] Explainable AI (XAI) in Computational Pathology Pipelines: Translating Machine Learning Features into Pathologist-Friendly Language
    Fine, Jeffrey
    Tosun, Akif
    Taylor, D. Lansing
    Becich, Michael
    Chennubhotla, S. Chakra
    LABORATORY INVESTIGATION, 2019, 99
  • [10] An Explainable Artificial Intelligence (xAI) Framework for Improving Trust in Automated ATM Tools
    Hernandez, Carolina Sanchez
    Ayo, Samuel
    Panagiotakopoulos, Dimitrios
    2021 IEEE/AIAA 40TH DIGITAL AVIONICS SYSTEMS CONFERENCE (DASC), 2021,