MGKGR: Multimodal Semantic Fusion for Geographic Knowledge Graph Representation

被引:0
|
作者
Zhang, Jianqiang [1 ]
Chen, Renyao [1 ]
Li, Shengwen [1 ,2 ,3 ]
Li, Tailong [4 ]
Yao, Hong [1 ,2 ,3 ,4 ]
机构
[1] China Univ Geosci, Sch Comp Sci, Wuhan 430074, Peoples R China
[2] China Univ Geosci, State Key Lab Biogeol & Environm Geol, Wuhan 430074, Peoples R China
[3] China Univ Geosci, Hubei Key Lab Intelligent Geoinformat Proc, Wuhan 430078, Peoples R China
[4] China Univ Geosci, Sch Future Technol, Wuhan 430074, Peoples R China
关键词
multimodal; geographic knowledge graph; knowledge graph representation;
D O I
10.3390/a17120593
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Geographic knowledge graph representation learning embeds entities and relationships in geographic knowledge graphs into a low-dimensional continuous vector space, which serves as a basic method that bridges geographic knowledge graphs and geographic applications. Previous geographic knowledge graph representation methods primarily learn the vectors of entities and their relationships from their spatial attributes and relationships, which ignores various semantics of entities, resulting in poor embeddings on geographic knowledge graphs. This study proposes a two-stage multimodal geographic knowledge graph representation (MGKGR) model that integrates multiple kinds of semantics to improve the embedding learning of geographic knowledge graph representation. Specifically, in the first stage, a spatial feature fusion method for modality enhancement is proposed to combine the structural features of geographic knowledge graphs with two modal semantic features. In the second stage, a multi-level modality feature fusion method is proposed to integrate heterogeneous features from different modalities. By fusing the semantics of text and images, the performance of geographic knowledge graph representation is improved, providing accurate representations for downstream geographic intelligence tasks. Extensive experiments on two datasets show that the proposed MGKGR model outperforms the baselines. Moreover, the results demonstrate that integrating textual and image data into geographic knowledge graphs can effectively enhance the model's performance.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] A Spatial-Spectral Bilinear Representation Fusion Network for Multimodal Classification
    Song, Xue
    Li, Lingling
    Jiao, Licheng
    Liu, Fang
    Liu, Xu
    Yang, Shuyuan
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61 : 1 - 17
  • [22] Relation-enhanced Negative Sampling for Multimodal Knowledge Graph Completion
    Xu, Derong
    Xu, Tong
    Wu, Shiwei
    Zhou, Jingbo
    Chen, Enhong
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 3857 - 3866
  • [23] FusionNN: A Semantic Feature Fusion Model Based on Multimodal for Web Anomaly Detection
    Wang, Li
    Xia, Mingshan
    Hu, Hao
    Li, Jianfang
    Hou, Fengyao
    Chen, Gang
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 79 (02): : 2991 - 3006
  • [24] DGFN Multimodal Emotion Analysis Model Based on Dynamic Graph Fusion Network
    Li, Jingwei
    Bai, Xinyi
    Han, Zhaoming
    INTERNATIONAL JOURNAL OF DECISION SUPPORT SYSTEM TECHNOLOGY, 2024, 16 (01)
  • [25] Source Code Vulnerability Detection Based on Joint Graph and Multimodal Feature Fusion
    Jin, Dun
    He, Chengwan
    Zou, Quan
    Qin, Yan
    Wang, Boshu
    ELECTRONICS, 2025, 14 (05):
  • [26] A Spatially Constraint Negative Sample Generation Method for Geographic Knowledge Graph Embedding
    Gao Y.
    Meng H.
    Ye C.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2023, 59 (03): : 434 - 444
  • [27] Is Visual Context Really Helpful for Knowledge Graph? A Representation Learning Perspective
    Wang, Meng
    Wang, Sen
    Yang, Han
    Zhang, Zheng
    Chen, Xi
    Qi, Guilin
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2735 - 2743
  • [28] Knowledge Graph Representation of Typhoon Disaster Events based on Spatiotemporal Processes
    Wang Y.
    Zhang X.
    Dang Y.
    Ye P.
    Journal of Geo-Information Science, 2023, 25 (06) : 1228 - 1239
  • [29] TIVA-KG: A Multimodal Knowledge Graph with Text, Image, Video and Audio
    Wang, Xin
    Meng, Benyuan
    Chen, Hong
    Meng, Yuan
    Lv, Ke
    Zhu, Wenwu
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 2391 - 2399
  • [30] Game-on: graph attention network based multimodal fusion for fake news detection
    Dhawan, Mudit
    Sharma, Shakshi
    Kadam, Aditya
    Sharma, Rajesh
    Kumaraguru, Ponnurangam
    SOCIAL NETWORK ANALYSIS AND MINING, 2024, 14 (01)