Multi-modal deep learning approaches to semantic segmentation of mining footprints with multispectral satellite imagery

被引:0
|
作者
Saputra, Muhamad Risqi U. [1 ]
Bhaswara, Irfan Dwiki [1 ]
Nasution, Bahrul Ilmi [1 ]
Ern, Michelle Ang Li [2 ]
Husna, Nur Laily Romadhotul [1 ]
Witra, Tahjudil [1 ]
Feliren, Vicky [1 ]
Owen, John R. [3 ]
Kemp, Deanna [4 ]
Lechner, Alex M. [1 ]
机构
[1] Monash Univ Indonesia, Min Spatial Data Intelligence Res Hub, Green Off Pk 9, Tangerang Selatan 15345, Banten, Indonesia
[2] Univ Nottingham Malaysia, Sch Environm & Geog Sci, Landscape Ecol & Conservat Lab, Semenyih 43500, Malaysia
[3] Univ Free State, Ctr Dev Support, 205 Nelson Mandela Dr,Pk West, ZA-9301 Bloemfontein, South Africa
[4] Univ Queensland, Sustainable Minerals Inst, Ctr Social Responsibil Min, Brisbane, Qld 4072, Australia
基金
澳大利亚研究理事会;
关键词
Semantic segmentation; Global mining footprints; Multispectral; Deep learning; IMPACTS;
D O I
10.1016/j.rse.2024.114584
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Existing remote sensing applications in mining are often of limited scope, typically mapping multiple mining land covers for a single mine or only mapping mining extents or a single feature (e.g., tailings dam) for multiple mines across a region. Many of these works have a narrow focus on specific mine land covers rather than encompassing the variety of mining and non-mining land use in a mine site. This study presents a pioneering effort in performing deep learning-based semantic segmentation of 37 mining locations worldwide, representing a range of commodities from gold to coal, using multispectral satellite imagery, to automate mapping of mining and non-mining land covers. Due to the absence of a dedicated training dataset, we crafted a customized multispectral dataset for training and testing deep learning models, leveraging and refining existing datasets in terms of boundaries, shapes, and class labels. We trained and tested multimodal semantic segmentation models, particularly based on U-Net, DeepLabV3+, Feature Pyramid Network (FPN), SegFormer, and IBM-NASA foundational geospatial model (Prithvi) architecture, with a focus on evaluating different model configurations, input band combinations, and the effectiveness of transfer learning. In terms of multimodality, we utilized various image bands, including Red, Green, Blue, and Near Infra-Red (NIR) and Normalized Difference Vegetation Index (NDVI), to determine which combination of inputs yields the most accurate segmentation. Results indicated that among different configurations, FPN with DenseNet-121 backbone, pre-trained on ImageNet, and trained using both RGB and NIR bands, performs the best. We concluded the study with a comprehensive assessment of the model's performance based on climate classification categories and diverse mining commodities. We believe that this work lays a robust foundation for further analysis of the complex relationship between mining projects, communities, and the environment.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Multi-modal neural machine translation with deep semantic interactions
    Su, Jinsong
    Chen, Jinchang
    Jiang, Hui
    Zhou, Chulun
    Lin, Huan
    Ge, Yubin
    Wu, Qingqiang
    Lai, Yongxuan
    INFORMATION SCIENCES, 2021, 554 : 47 - 60
  • [42] Multi-modal NeRF Self-Supervision for LiDAR Semantic Segmentation
    Timoneda, Xavier
    Herb, Markus
    Duerr, Fabian
    Goehring, Daniel
    Yu, Fisher
    2024 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS 2024), 2024, : 12939 - 12946
  • [43] EISNet: A Multi-Modal Fusion Network for Semantic Segmentation With Events and Images
    Xie, Bochen
    Deng, Yongjian
    Shao, Zhanpeng
    Li, Youfu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 8639 - 8650
  • [44] Monitoring deforestation in Jordan using deep semantic segmentation with satellite imagery
    Alzu'bi, Ahmad
    Alsmadi, Lujain
    ECOLOGICAL INFORMATICS, 2022, 70
  • [45] Farmland Recognition of High Resolution Multispectral Remote Sensing Imagery using Deep Learning Semantic Segmentation Method
    Shuangpeng, Zheng
    Ta, Fang
    Hong, Huo
    PROCEEDINGS OF 2019 INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE (PRAI 2019), 2019, : 33 - 40
  • [46] End-to-end Cloud Segmentation in High-Resolution Multispectral Satellite Imagery Using Deep Learning
    Morales, Giorgio
    Ramirez, Alejandro
    Telles, Joel
    PROCEEDINGS OF THE 2019 IEEE XXVI INTERNATIONAL CONFERENCE ON ELECTRONICS, ELECTRICAL ENGINEERING AND COMPUTING (INTERCON), 2019,
  • [47] A Transformer-based multi-modal fusion network for semantic segmentation of high-resolution remote sensing imagery
    Liu, Yutong
    Gao, Kun
    Wang, Hong
    Yang, Zhijia
    Wang, Pengyu
    Ji, Shijing
    Huang, Yanjun
    Zhu, Zhenyu
    Zhao, Xiaobin
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2024, 133
  • [48] Segmentation of Multi-Modal MRI Brain Tumor Sub-Regions Using Deep Learning
    B. Srinivas
    Gottapu Sasibhushana Rao
    Journal of Electrical Engineering & Technology, 2020, 15 : 1899 - 1909
  • [49] TSC-DL: Unsupervised Trajectory Segmentation of Multi-Modal Surgical Demonstrations with Deep Learning
    Murali, Adithyavairavan
    Garg, Animesh
    Krishnan, Sanjay
    Pokorny, Florian T.
    Abbeel, Pieter
    Darrell, Trevor
    Goldberg, Ken
    2016 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2016, : 4150 - 4157
  • [50] Multi-Modal Song Mood Detection with Deep Learning
    Pyrovolakis, Konstantinos
    Tzouveli, Paraskevi
    Stamou, Giorgos
    SENSORS, 2022, 22 (03)