Deep Learning for HABs Prediction with Multimodal Fusion

被引:0
|
作者
Zhao, Fei [1 ]
Zhang, Chengcui [1 ]
机构
[1] Univ Alabama Birmingham, Birmingham, AL 35294 USA
来源
31ST ACM SIGSPATIAL INTERNATIONAL CONFERENCE ON ADVANCES IN GEOGRAPHIC INFORMATION SYSTEMS, ACM SIGSPATIAL GIS 2023 | 2023年
关键词
Geolocation; Computer Vision; Deep Learning; Harmful Algal Blooms;
D O I
10.1145/3589132.3628370
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Harmful Algal Blooms (HABs) present significant environmental and public health threats. Recent machine learning-based HABs monitoring methods often rely solely on unimodal data, e.g., satellite imagery, overlooking crucial environmental factors such as temperature. Moreover, existing multi-modal approaches grapple with real-time applicability and generalizability challenges due to the use of ensemble methodologies and hard-coded geolocation clusters. Addressing these gaps, this paper presents a novel deep learning model using a single-model-based multi-task framework. This framework is designed to segment water bodies and predict HABs severity levels concurrently, enabling the model to focus on areas of interest, thereby enhancing prediction accuracy. Our model integrates multimodal inputs, i.e., satellite imagery, elevation data, temperature readings, and geolocation details, via a dual-branch architecture: the Satellite-Elevation (SE) branch and the TemperatureGeolocation (TG) branch. Satellite and elevation data in the SE branch, being spatially coherent, assist in water area detection and feature extraction. Meanwhile, the TG branch, using sequential temperature data and geolocation information, captures temporal algal growth patterns and adjusts for temperature variations influenced by regional climatic differences, ensuring the model's adaptability across different geographic regions. Additionally, we propose a geometric multimodal focal loss to further enhance representation learning. On the Tick-Tick Bloom (TTB) dataset, our approach outperforms the SOTA methods by 15.65%.
引用
收藏
页码:17 / 18
页数:2
相关论文
共 50 条
  • [1] Deep Multimodal Fusion for Persuasiveness Prediction
    Nojavanasghari, Behnaz
    Gopinath, Deepak
    Koushik, Jayanth
    Baltrusaitis, Tadas
    Morency, Louis-Philippe
    ICMI'16: PROCEEDINGS OF THE 18TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2016, : 284 - 288
  • [2] A Survey on Deep Learning for Multimodal Data Fusion
    Gao, Jing
    Li, Peng
    Chen, Zhikui
    Zhang, Jianing
    NEURAL COMPUTATION, 2020, 32 (05) : 829 - 864
  • [3] Advances in deep learning for multimodal fusion and alignment
    Multimedia Tools and Applications, 2022, 81 : 11931 - 11931
  • [4] Advances in deep learning for multimodal fusion and alignment
    Huang, Feiran
    Mumtaz, Shahid
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (09) : 11931 - 11931
  • [5] Soybean yield prediction from UAV using multimodal data fusion and deep learning
    Maimaitijiang, Maitiniyazi
    Sagan, Vasit
    Sidike, Paheding
    Hartling, Sean
    Esposito, Flavin
    Fritschi, Felix B.
    REMOTE SENSING OF ENVIRONMENT, 2020, 237
  • [6] Exploring Fusion Strategies in Deep Multimodal Affect Prediction
    Patania, Sabrina
    D'Amelio, Alessandro
    Lanzarotti, Raffaella
    IMAGE ANALYSIS AND PROCESSING, ICIAP 2022, PT II, 2022, 13232 : 730 - 741
  • [7] Visualizing Multimodal Deep Learning for Lesion Prediction
    Gillmann, Christina
    Peter, Lucas
    Schmidt, Carlo
    Saur, Dorothee
    Scheuermann, Gerik
    IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2021, 41 (05) : 90 - 98
  • [8] Success Prediction on Crowdfunding with Multimodal Deep Learning
    Cheng, Chaoran
    Tan, Fei
    Hou, Xiurui
    Wei, Zhi
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2158 - 2164
  • [9] Multimodal Deep Learning for Solar Irradiance Prediction
    Li, Zhuo
    Wang, Kejie
    Li, Chenchen
    Zhao, Miao
    Cao, Jiannong
    2019 INTERNATIONAL CONFERENCE ON INTERNET OF THINGS (ITHINGS) AND IEEE GREEN COMPUTING AND COMMUNICATIONS (GREENCOM) AND IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING (CPSCOM) AND IEEE SMART DATA (SMARTDATA), 2019, : 784 - 792
  • [10] Deep Learning with Multimodal Data for Prediction of Glaucoma
    Chen, Jimmy
    Fan, Rui
    Christopher, Mark
    Belghith, Akram
    Bowd, Christopher
    Goldbaum, Michael
    Fazio, Massimo
    Girkin, Christopher
    De Moraes, Gustavo
    Liebmann, Jeffrey
    Zangwill, Linda
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2023, 64 (09)