Explainable artificial intelligence in geoscience: A glimpse into the future of landslide susceptibility modeling

被引:49
作者
Dahal, Ashok [1 ]
Lombardo, Luigi [1 ]
机构
[1] Univ Twente, Fac Geoinformat Sci & Earth Observ ITC, POB 217, NL-7500 AE Enschede, Netherlands
关键词
Landslide modeling; Explainable deep learning; Nepal Earthquake; Web-GIS; Transparent modeling; LOGISTIC-REGRESSION; NEURAL-NETWORKS; QUANTITATIVE-ANALYSIS; HAZARD; REGION; NORTH; INFORMATION; PERFORMANCE; TECHNOLOGY; VALIDATION;
D O I
10.1016/j.cageo.2023.105364
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
For decades, the distinction between statistical models and machine learning ones has been clear. The former are optimized to produce interpretable results, whereas the latter seeks to maximize the predictive performance of the task at hand. This is valid for any scientific field and for any method belonging to the two categories mentioned above. When attempting to predict natural hazards, this difference has lead researchers to make drastic decisions on which aspect to prioritize, a difficult choice to make. In fact, one would always seek the highest performance because at higher performances correspond better decisions for disaster risk reduction. However, scientists also wish to understand the results, as a way to rely on the tool they developed. Today, very recent development in deep learning have brought forward a new generation of interpretable artificial intelli-gence, where the prediction power typical of machine learning tools is equipped with a level of explanatory power typical of statistical approaches. In this work, we attempt to demonstrate the capabilities of this new generation of explainable artificial intelligence (XAI). To do so, we take the landslide susceptibility context as reference. Specifically, we build an XAI trained to model landslides occurred in response to the Gorkha earth-quake (April 25, 2015), providing an educational overview of the model design and its querying opportunities. The results show high performance, with an AUC score of 0.89, while the interpretability can be extended to the probabilistic result assigned to single mapping units.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Landslide susceptibility modeling by interpretable neural network
    Youssef, K.
    Shao, K.
    Moon, S.
    Bouchard, L. -s.
    COMMUNICATIONS EARTH & ENVIRONMENT, 2023, 4 (01):
  • [22] Double landslide susceptibility assessment based on artificial neural networks and weights of evidence
    Goyes-Penafiel, Paul
    Hernandez-Rojas, Alejandra
    BOLETIN DE GEOLOGIA, 2021, 43 (01): : 173 - 191
  • [23] Shallow Landslide Susceptibility Modeling Using the Data Mining Models Artificial Neural Network and Boosted Tree
    Oh, Hyun-Joo
    Lee, Saro
    APPLIED SCIENCES-BASEL, 2017, 7 (10):
  • [24] Interpretable machine learning and explainable artificial intelligence
    Topuz, Kazim
    Bajaj, Akhilesh
    Coussement, Kristof
    Urban, Timothy L.
    ANNALS OF OPERATIONS RESEARCH, 2025, : 775 - 782
  • [25] Explainable Artificial Intelligence (XAI) Adoption and Advocacy
    Ridley, Michael
    INFORMATION TECHNOLOGY AND LIBRARIES, 2022, 41 (02)
  • [26] Explainable Artificial Intelligence for Tabular Data: A Survey
    Sahakyan, Maria
    Aung, Zeyar
    Rahwan, Talal
    IEEE ACCESS, 2021, 9 : 135392 - 135422
  • [27] The challenges of integrating explainable artificial intelligence into GeoAI
    Xing, Jin
    Sieber, Renee
    TRANSACTIONS IN GIS, 2023, 27 (03) : 626 - 645
  • [28] Artificial intelligence: Implications for the future of work
    Howard, John
    AMERICAN JOURNAL OF INDUSTRIAL MEDICINE, 2019, 62 (11) : 917 - 926
  • [29] Artificial Intelligence in Surgery: The Future Is Now
    Guni, Ahmad
    Varma, Piyush
    Zhang, Joe
    Fehervari, Matyas
    Ashrafian, Hutan
    EUROPEAN SURGICAL RESEARCH, 2024, 65 (01) : 22 - 39
  • [30] Artificial neural network ensembles applied to the mapping of landslide susceptibility
    Bragagnolo, L.
    da Silva, R., V
    Grzybowski, J. M., V
    CATENA, 2020, 184