Representation and Fusion Based on Knowledge Graph in Multi-Modal Semantic Communication

被引:2
|
作者
Xing, Chenlin [1 ]
Lv, Jie [1 ]
Luo, Tao [1 ]
Zhang, Zhilong [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Informat & Commun Engn, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Correlation; Feature extraction; Knowledge graphs; Cognition; Head; Data mining; Semantic communication; multi-modal fusion; knowledge graph;
D O I
10.1109/LWC.2024.3369864
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The existing research on multi-modal semantic communication ignores the exploration of reasoning correlation among multi-modal data. Motivated by this, a multi-modal semantic representation and fusion model based on knowledge graph (KG-MSF) is proposed in this letter. In KG-MSF, the direct and reasoning correlation semantic information is extracted and mapped into a two-layer semantic architecture to represent the semantics of each modal fully. After that, the knowledge graph with structural advantage is utilized to fuse multi-modal semantic information, which is transmitted under different channel conditions. To assess the efficacy of semantic representation and fusion of the proposed KG-MSF in the multi-modal semantic communication system, we conduct comprehensive experiments on the task of visual question answer (VQA) with a metric of answer accuracy. The results demonstrate the superiority compared with existing models for multi-modal semantic representation, fusion, transmission efficiency and channel robustness.
引用
收藏
页码:1344 / 1348
页数:5
相关论文
共 50 条
  • [21] Oracle Bone Inscriptions information processing based on multi-modal knowledge graph
    Xiong, Jing
    Liu, Guoying
    Liu, Yongge
    Liu, Mengting
    COMPUTERS & ELECTRICAL ENGINEERING, 2021, 92
  • [22] A Discriminative Vectorial Framework for Multi-Modal Feature Representation
    Gao, Lei
    Guan, Ling
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 1503 - 1514
  • [23] Medical Visual Question-Answering Model Based on Knowledge Enhancement and Multi-Modal Fusion
    Zhang, Dianyuan
    Yu, Chuanming
    An, Lu
    Proceedings of the Association for Information Science and Technology, 2024, 61 (01) : 703 - 708
  • [24] AutoCite: Multi-Modal Representation Fusion for Contextual Citation Generation
    Wang, Qingqin
    Xiong, Yun
    Zhang, Yao
    Zhang, Jiawei
    Zhu, Yangyong
    WSDM '21: PROCEEDINGS OF THE 14TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2021, : 788 - 796
  • [25] Robust 3D Semantic Segmentation Based on Multi-Phase Multi-Modal Fusion for Intelligent Vehicles
    Ni, Peizhou
    Li, Xu
    Xu, Wang
    Kong, Dong
    Hu, Yue
    Wei, Kun
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2024, 9 (01): : 1602 - 1614
  • [26] Multi-modal Graph Learning over UMLS Knowledge Graphs
    Burger, Manuel
    Ratsch, Gunnar
    Kuznetsova, Rita
    MACHINE LEARNING FOR HEALTH, ML4H, VOL 225, 2023, 225 : 52 - 81
  • [27] Cognitive knowledge graph generation for grid fault handling based on attention mechanism combined with multi-modal factor fusion
    Li, Zhenbin
    Huang, Zhigang
    Guo, Lingxu
    Shan, Lianfei
    Yu, Guangyao
    Chong, Zhiqiang
    Zhang, Yue
    COMPUTERS & ELECTRICAL ENGINEERING, 2023, 111
  • [28] MMpedia: A Large-Scale Multi-modal Knowledge Graph
    Wu, Yinan
    Wu, Xiaowei
    Li, Junwen
    Zhang, Yue
    Wang, Haofen
    Du, Wen
    He, Zhidong
    Liu, Jingping
    Ruan, Tong
    SEMANTIC WEB, ISWC 2023, PT II, 2023, 14266 : 18 - 37
  • [29] A Survey of Knowledge Graph Reasoning on Graph Types: Static, Dynamic, and Multi-Modal
    Liang, Ke
    Meng, Lingyuan
    Liu, Meng
    Liu, Yue
    Tu, Wenxuan
    Wang, Siwei
    Zhou, Sihang
    Liu, Xinwang
    Sun, Fuchun
    He, Kunlun
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 9456 - 9478
  • [30] Multi-modal fusion architecture search for camera-based semantic scene completion
    Wang, Xuzhi
    Feng, Wei
    Wan, Liang
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 243