On the Uses of Large Language Models to Design End-to-end Learning Semantic Communication

被引:0
|
作者
Wang, Ying [1 ]
Sun, Zhuo [1 ]
Fan, Jinpo [1 ]
Ma, Hao [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Key Lab Univ Wireless Commun, Minist Educ, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
end-to-end learning; large language models; semantic communication;
D O I
10.1109/WCNC57260.2024.10570717
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning-based semantic communication is a promising research direction for next-generation communication systems. The emergence of large language models(LLMs) with remarkable semantic comprehension abilities leads us to consider whether LLMs can be used in semantic communication to enhance model's performance. In this paper, we discuss the main implementing details of the idea by proposing a general end-to-end learning semantic communication model with LLM, including subword-level tokenization, a rate adapter based on gradients for matching the rate requirements of any channel codec and fine-tuning for possessing private background knowledge. By taking Bidirectional and Auto-Regressive Transformers (BART) and Generative Pre-trained Transformer 2 (GPT2) as examples, we demonstrate how we can utilize various structures of LLMs to design semantic codecs. In terms of semantic fidelity, generalizability to cross-scenario, and complexity, results reveal that the LLM-based semantic communication system achieves exciting performance. We hope this initial work can inspire more research devoted to this field.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Exploring Transfer Learning For End-to-End Spoken Language Understanding
    Rongali, Subendhu
    Liu, Beiye
    Cai, Liwei
    Arkoudas, Konstantine
    Su, Chengwei
    Hamza, Wael
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 13754 - 13761
  • [22] End-to-End Learning of Communication System without Known Channel
    Jiang, Hao
    Dai, Linglong
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [23] POLYNOMIAL END-TO-END COMMUNICATION
    AWERBUCH, B
    MANSOUR, Y
    SHAVIT, N
    30TH ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, 1989, : 358 - 363
  • [24] Correction to: An end-to-end differential network learning method for semantic segmentation
    Tai Hu
    Ming Yang
    Wanqi Yang
    Aishi Li
    International Journal of Machine Learning and Cybernetics, 2019, 10 : 1925 - 1925
  • [25] End-to-End Learning for RIS-Aided Communication Systems
    Jiang, Hao
    Dai, Linglong
    Hao, Mo
    MacKenzie, Richard
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (06) : 6778 - 6783
  • [26] End-to-end Convolutional Semantic Embeddings
    You, Quanzeng
    Zhang, Zhengyou
    Luo, Jiebo
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 5735 - 5744
  • [27] A framework for end-to-end learning on semantic tree-structured data
    Woof, William
    Chen, Ke
    arXiv, 2020,
  • [28] End-to-End Emotion Semantic Parsing
    Jiang, Xiaotong
    Wang, Zhongqing
    Zhou, Guodong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 37 - 47
  • [29] Codebook-enabled Generative End-to-end Semantic Communication Powered by Transformer
    Ye, Peigen
    Sun, Yaping
    Yao, Shumin
    Chen, Hao
    Xu, Xiaodong
    Cui, Shuguang
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS, INFOCOM WKSHPS 2024, 2024,
  • [30] Building Semantic Communication System via Molecules: An End-to-End Training Approach
    Cheng, Yukun
    Wei, Chen
    Bo, Ai
    CHINA COMMUNICATIONS, 2024, 21 (07) : 113 - 124