Multi-modal deep learning approaches to semantic segmentation of mining footprints with multispectral satellite imagery

被引:0
|
作者
Saputra, Muhamad Risqi U. [1 ]
Bhaswara, Irfan Dwiki [1 ]
Nasution, Bahrul Ilmi [1 ]
Ern, Michelle Ang Li [2 ]
Husna, Nur Laily Romadhotul [1 ]
Witra, Tahjudil [1 ]
Feliren, Vicky [1 ]
Owen, John R. [3 ]
Kemp, Deanna [4 ]
Lechner, Alex M. [1 ]
机构
[1] Monash Univ Indonesia, Min Spatial Data Intelligence Res Hub, Green Off Pk 9, Tangerang Selatan 15345, Banten, Indonesia
[2] Univ Nottingham Malaysia, Sch Environm & Geog Sci, Landscape Ecol & Conservat Lab, Semenyih 43500, Malaysia
[3] Univ Free State, Ctr Dev Support, 205 Nelson Mandela Dr,Pk West, ZA-9301 Bloemfontein, South Africa
[4] Univ Queensland, Sustainable Minerals Inst, Ctr Social Responsibil Min, Brisbane, Qld 4072, Australia
基金
澳大利亚研究理事会;
关键词
Semantic segmentation; Global mining footprints; Multispectral; Deep learning; IMPACTS;
D O I
10.1016/j.rse.2024.114584
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Existing remote sensing applications in mining are often of limited scope, typically mapping multiple mining land covers for a single mine or only mapping mining extents or a single feature (e.g., tailings dam) for multiple mines across a region. Many of these works have a narrow focus on specific mine land covers rather than encompassing the variety of mining and non-mining land use in a mine site. This study presents a pioneering effort in performing deep learning-based semantic segmentation of 37 mining locations worldwide, representing a range of commodities from gold to coal, using multispectral satellite imagery, to automate mapping of mining and non-mining land covers. Due to the absence of a dedicated training dataset, we crafted a customized multispectral dataset for training and testing deep learning models, leveraging and refining existing datasets in terms of boundaries, shapes, and class labels. We trained and tested multimodal semantic segmentation models, particularly based on U-Net, DeepLabV3+, Feature Pyramid Network (FPN), SegFormer, and IBM-NASA foundational geospatial model (Prithvi) architecture, with a focus on evaluating different model configurations, input band combinations, and the effectiveness of transfer learning. In terms of multimodality, we utilized various image bands, including Red, Green, Blue, and Near Infra-Red (NIR) and Normalized Difference Vegetation Index (NDVI), to determine which combination of inputs yields the most accurate segmentation. Results indicated that among different configurations, FPN with DenseNet-121 backbone, pre-trained on ImageNet, and trained using both RGB and NIR bands, performs the best. We concluded the study with a comprehensive assessment of the model's performance based on climate classification categories and diverse mining commodities. We believe that this work lays a robust foundation for further analysis of the complex relationship between mining projects, communities, and the environment.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] Multi-modal Prototypes for Open-World Semantic Segmentation
    Yang, Yuhuan
    Ma, Chaofan
    Ju, Chen
    Zhang, Fei
    Yao, Jiangchao
    Zhang, Ya
    Wang, Yanfeng
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (12) : 6004 - 6020
  • [22] Semantic Segmentation of Defects in Infrastructures through Multi-modal Images
    Shahsavarani, Sara
    Lopez, Fernando
    Ibarra-Castanedo, Clemente
    Maldague, Xavier P., V
    THERMOSENSE: THERMAL INFRARED APPLICATIONS XLVI, 2024, 13047
  • [23] Ticino: A multi-modal remote sensing dataset for semantic segmentation
    Barbato, Mirko Paolo
    Piccoli, Flavio
    Napoletano, Paolo
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 249
  • [24] Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges
    Feng, Di
    Haase-Schutz, Christian
    Rosenbaum, Lars
    Hertlein, Heinz
    Glaser, Claudius
    Timm, Fabian
    Wiesbeck, Werner
    Dietmayer, Klaus
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (03) : 1341 - 1360
  • [25] MULTI-MODAL MULTI-TASK LEARNING FOR SEMANTIC SEGMENTATION OF LAND COVER UNDER CLOUDY CONDITIONS
    Xu, Fang
    Shi, Yilei
    Yang, Wen
    Zhu, Xiaoxiang
    IGARSS 2023 - 2023 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2023, : 6274 - 6277
  • [26] Deep learning approaches for multi-modal sensor data analysis and abnormality detection
    Jadhav, Santosh Pandurang
    Srinivas, Angalkuditi
    Dipak Raghunath, Patil
    Ramkumar Prabhu, M.
    Suryawanshi, Jaya
    Haldorai, Anandakumar
    Measurement: Sensors, 33
  • [27] Multi-task Learning of Semantic Segmentation and Height Estimation for Multi-modal Remote Sensing Images
    Mengyu WANG
    Zhiyuan YAN
    Yingchao FENG
    Wenhui DIAO
    Xian SUN
    Journal of Geodesy and Geoinformation Science, 2023, 6 (04) : 27 - 39
  • [28] An explainable deep learning pipeline for multi-modal multi-organ medical image segmentation
    Mylona, E.
    Zaridis, D.
    Grigoriadis, G.
    Tachos, N.
    Fotiadis, D. I.
    RADIOTHERAPY AND ONCOLOGY, 2022, 170 : S275 - S276
  • [29] Deep Learning Approach for Building Detection in Satellite Multispectral Imagery
    Prathap, Geesara
    Afanasyev, Ilya
    2018 9TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS (IS), 2018, : 461 - 465
  • [30] Multi-modal deep learning for landform recognition
    Du, Lin
    You, Xiong
    Li, Ke
    Meng, Liqiu
    Cheng, Gong
    Xiong, Liyang
    Wang, Guangxia
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2019, 158 : 63 - 75