Multi-modal deep learning approaches to semantic segmentation of mining footprints with multispectral satellite imagery

被引:0
|
作者
Saputra, Muhamad Risqi U. [1 ]
Bhaswara, Irfan Dwiki [1 ]
Nasution, Bahrul Ilmi [1 ]
Ern, Michelle Ang Li [2 ]
Husna, Nur Laily Romadhotul [1 ]
Witra, Tahjudil [1 ]
Feliren, Vicky [1 ]
Owen, John R. [3 ]
Kemp, Deanna [4 ]
Lechner, Alex M. [1 ]
机构
[1] Monash Univ Indonesia, Min Spatial Data Intelligence Res Hub, Green Off Pk 9, Tangerang Selatan 15345, Banten, Indonesia
[2] Univ Nottingham Malaysia, Sch Environm & Geog Sci, Landscape Ecol & Conservat Lab, Semenyih 43500, Malaysia
[3] Univ Free State, Ctr Dev Support, 205 Nelson Mandela Dr,Pk West, ZA-9301 Bloemfontein, South Africa
[4] Univ Queensland, Sustainable Minerals Inst, Ctr Social Responsibil Min, Brisbane, Qld 4072, Australia
基金
澳大利亚研究理事会;
关键词
Semantic segmentation; Global mining footprints; Multispectral; Deep learning; IMPACTS;
D O I
10.1016/j.rse.2024.114584
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Existing remote sensing applications in mining are often of limited scope, typically mapping multiple mining land covers for a single mine or only mapping mining extents or a single feature (e.g., tailings dam) for multiple mines across a region. Many of these works have a narrow focus on specific mine land covers rather than encompassing the variety of mining and non-mining land use in a mine site. This study presents a pioneering effort in performing deep learning-based semantic segmentation of 37 mining locations worldwide, representing a range of commodities from gold to coal, using multispectral satellite imagery, to automate mapping of mining and non-mining land covers. Due to the absence of a dedicated training dataset, we crafted a customized multispectral dataset for training and testing deep learning models, leveraging and refining existing datasets in terms of boundaries, shapes, and class labels. We trained and tested multimodal semantic segmentation models, particularly based on U-Net, DeepLabV3+, Feature Pyramid Network (FPN), SegFormer, and IBM-NASA foundational geospatial model (Prithvi) architecture, with a focus on evaluating different model configurations, input band combinations, and the effectiveness of transfer learning. In terms of multimodality, we utilized various image bands, including Red, Green, Blue, and Near Infra-Red (NIR) and Normalized Difference Vegetation Index (NDVI), to determine which combination of inputs yields the most accurate segmentation. Results indicated that among different configurations, FPN with DenseNet-121 backbone, pre-trained on ImageNet, and trained using both RGB and NIR bands, performs the best. We concluded the study with a comprehensive assessment of the model's performance based on climate classification categories and diverse mining commodities. We believe that this work lays a robust foundation for further analysis of the complex relationship between mining projects, communities, and the environment.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] SEMANTIC SEGMENTATION OF UNDERWATER SONAR IMAGERY WITH DEEP LEARNING
    Rahnemoonfar, Maryam
    Dobbs, Dugan
    2019 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2019), 2019, : 9455 - 9458
  • [32] Deep Multi-Modal Vehicle Detection in Aerial ISR Imagery
    Sakla, Wesam
    Konjevod, Goran
    Mundhenk, T. Nathan
    2017 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2017), 2017, : 916 - 923
  • [33] Deep Multi-modal Learning with Cascade Consensus
    Yang, Yang
    Wu, Yi-Feng
    Zhan, De-Chuan
    Jiang, Yuan
    PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT II, 2018, 11013 : 64 - 72
  • [34] Multi-modal deep distance metric learning
    Roostaiyan, Seyed Mahdi
    Imani, Ehsan
    Baghshah, Mahdieh Soleymani
    INTELLIGENT DATA ANALYSIS, 2017, 21 (06) : 1351 - 1369
  • [35] Robust 3D Semantic Segmentation Method Based on Multi-Modal Collaborative Learning
    Ni, Peizhou
    Li, Xu
    Xu, Wang
    Zhou, Xiaojing
    Jiang, Tao
    Hu, Weiming
    REMOTE SENSING, 2024, 16 (03)
  • [36] Toward Unpaired Multi-modal Medical Image Segmentation via Learning Structured Semantic Consistency
    Yang, Jie
    Zhu, Ye
    Wang, Chaoqun
    Li, Zhen
    Zhang, Ruimao
    MEDICAL IMAGING WITH DEEP LEARNING, VOL 227, 2023, 227 : 1602 - 1622
  • [37] Adversarial unsupervised domain adaptation for 3D semantic segmentation with multi-modal learning
    Liu, Wei
    Luo, Zhiming
    Cai, Yuanzheng
    Yu, Ying
    Ke, Yang
    Marcato Junior, Jose
    Goncalves, Wesley Nunes
    Li, Jonathan
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2021, 176 : 211 - 221
  • [38] Semantic segmentation of multispectral photoacoustic images using deep learning
    Schellenberg, Melanie
    Dreher, Kris K.
    Holzwarth, Niklas
    Isensee, Fabian
    Reinke, Annika
    Schreck, Nicholas
    Seitel, Alexander
    Tizabi, Minu D.
    Maier-Hein, Lena
    Groehl, Janek
    PHOTOACOUSTICS, 2022, 26
  • [39] A multi-modal dental dataset for semi-supervised deep learning image segmentation
    Wang, Yaqi
    Ye, Fan
    Chen, Yifei
    Wang, Chengkai
    Wu, Chengyu
    Xu, Feng
    Ma, Zhean
    Liu, Yi
    Zhang, Yifan
    Cao, Mingguo
    Chen, Xiaodiao
    SCIENTIFIC DATA, 2025, 12 (01)
  • [40] MULTI-CLASS SEGMENTATION OF URBAN FLOODS FROM MULTISPECTRAL IMAGERY USING DEEP LEARNING
    Potnis, Abhishek V.
    Shinde, Rajat C.
    Durbha, Surya S.
    Kurte, Kuldeep R.
    2019 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2019), 2019, : 9741 - 9744