Representation and Fusion Based on Knowledge Graph in Multi-Modal Semantic Communication

被引:2
作者
Xing, Chenlin [1 ]
Lv, Jie [1 ]
Luo, Tao [1 ]
Zhang, Zhilong [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Informat & Commun Engn, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Correlation; Feature extraction; Knowledge graphs; Cognition; Head; Data mining; Semantic communication; multi-modal fusion; knowledge graph;
D O I
10.1109/LWC.2024.3369864
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The existing research on multi-modal semantic communication ignores the exploration of reasoning correlation among multi-modal data. Motivated by this, a multi-modal semantic representation and fusion model based on knowledge graph (KG-MSF) is proposed in this letter. In KG-MSF, the direct and reasoning correlation semantic information is extracted and mapped into a two-layer semantic architecture to represent the semantics of each modal fully. After that, the knowledge graph with structural advantage is utilized to fuse multi-modal semantic information, which is transmitted under different channel conditions. To assess the efficacy of semantic representation and fusion of the proposed KG-MSF in the multi-modal semantic communication system, we conduct comprehensive experiments on the task of visual question answer (VQA) with a metric of answer accuracy. The results demonstrate the superiority compared with existing models for multi-modal semantic representation, fusion, transmission efficiency and channel robustness.
引用
收藏
页码:1344 / 1348
页数:5
相关论文
共 50 条
  • [41] AF: An Association-Based Fusion Method for Multi-Modal Classification
    Liang, Xinyan
    Qian, Yuhua
    Guo, Qian
    Cheng, Honghong
    Liang, Jiye
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (12) : 9236 - 9254
  • [42] An Abnormal External Link Detection Algorithm Based on Multi-Modal Fusion
    Wu, Zhiqiang
    INTERNATIONAL JOURNAL OF INFORMATION SECURITY AND PRIVACY, 2024, 18 (01)
  • [43] Human activity recognition based on multi-modal fusion
    Cheng Zhang
    Tianqi Zu
    Yibin Hou
    Jian He
    Shengqi Yang
    Ruihai Dong
    CCF Transactions on Pervasive Computing and Interaction, 2023, 5 : 321 - 332
  • [44] Knowledge Synergy Learning for Multi-Modal Tracking
    He, Yuhang
    Ma, Zhiheng
    Wei, Xing
    Gong, Yihong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (07) : 5519 - 5532
  • [45] Single-Stage Extensive Semantic Fusion for multi-modal sarcasm detection
    Fang, Hong
    Liang, Dahao
    Xiang, Weiyu
    ARRAY, 2024, 22
  • [46] On Graph Calculi for Multi-modal Logics
    Veloso, Paulo A. S.
    Veloso, Sheila R. M.
    Benevides, Mario R. F.
    ELECTRONIC NOTES IN THEORETICAL COMPUTER SCIENCE, 2015, 312 : 231 - 252
  • [47] MHR: A Multi-Modal Hyperbolic Representation Framework for Fake News Detection
    Feng, Shanshan
    Yu, Guoxin
    Liu, Dawei
    Hu, Han
    Luo, Yong
    Lin, Hui
    Ong, Yew-Soon
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2025, 37 (04) : 2015 - 2028
  • [48] Fusion of Multi-Modal Underwater Ship Inspection Data with Knowledge Graphs
    Hirsch, Joseph
    Elvesaeter, Brian
    Cardaillac, Alexandre
    Bauer, Bernhard
    Waszak, Maryna
    2022 OCEANS HAMPTON ROADS, 2022,
  • [49] Robust multi-modal fusion architecture for medical data with knowledge distillation
    Wang, Muyu
    Fan, Shiyu
    Li, Yichen
    Gao, Binyu
    Xie, Zhongrang
    Chen, Hui
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2025, 260
  • [50] Human activity recognition based on multi-modal fusion
    Zhang, Cheng
    Zu, Tianqi
    Hou, Yibin
    He, Jian
    Yang, Shengqi
    Dong, Ruihai
    CCF TRANSACTIONS ON PERVASIVE COMPUTING AND INTERACTION, 2023, 5 (03) : 321 - 332