On the Uses of Large Language Models to Design End-to-end Learning Semantic Communication

被引:0
|
作者
Wang, Ying [1 ]
Sun, Zhuo [1 ]
Fan, Jinpo [1 ]
Ma, Hao [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Key Lab Univ Wireless Commun, Minist Educ, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
end-to-end learning; large language models; semantic communication;
D O I
10.1109/WCNC57260.2024.10570717
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning-based semantic communication is a promising research direction for next-generation communication systems. The emergence of large language models(LLMs) with remarkable semantic comprehension abilities leads us to consider whether LLMs can be used in semantic communication to enhance model's performance. In this paper, we discuss the main implementing details of the idea by proposing a general end-to-end learning semantic communication model with LLM, including subword-level tokenization, a rate adapter based on gradients for matching the rate requirements of any channel codec and fine-tuning for possessing private background knowledge. By taking Bidirectional and Auto-Regressive Transformers (BART) and Generative Pre-trained Transformer 2 (GPT2) as examples, we demonstrate how we can utilize various structures of LLMs to design semantic codecs. In terms of semantic fidelity, generalizability to cross-scenario, and complexity, results reveal that the LLM-based semantic communication system achieves exciting performance. We hope this initial work can inspire more research devoted to this field.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Building Semantic Communication System via Molecules: An End-to-End Training Approach
    Cheng Yukun
    Chen Wei
    Ai Bo
    ChinaCommunications, 2024, 21 (07) : 113 - 124
  • [32] Deep Learning in Physical Layer: Review on Data Driven End-to-End Communication Systems and Their Enabling Semantic Applications
    Islam, Nazmul
    Shin, Seokjoo
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2024, 5 : 4207 - 4240
  • [33] SEQ2SEQ-SC: END-TO-END SEMANTIC COMMUNICATION SYSTEMS WITH PRE-TRAINED LANGUAGE MODEL
    Lee, Ju-Hyung
    Lee, Dong-Ho
    Sheen, Eunsoo
    Choi, Thomas
    Pujara, Jay
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 260 - 264
  • [34] KNOWLEDGE TRANSFER FROM LARGE-SCALE PRETRAINED LANGUAGE MODELS TO END-TO-END SPEECH RECOGNIZERS
    Kubo, Yotaro
    Karita, Shigeki
    Bacchiani, Michiel
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8512 - 8516
  • [35] Using Large Language Model for End-to-End Chinese ASR and NER
    Li, Yuang
    Yu, Jiawei
    Zhang, Min
    Ren, Mengxin
    Zhao, Yanqing
    Zhao, Xiaofeng
    Tao, Shimin
    Su, Jinsong
    Yang, Hao
    INTERSPEECH 2024, 2024, : 822 - 826
  • [36] Online Continual Learning of End-to-End Speech Recognition Models
    Yang, Muqiao
    Lane, Ian
    Watanabe, Shinji
    INTERSPEECH 2022, 2022, : 2668 - 2672
  • [37] IMPROVING END-TO-END MODELS FOR SET PREDICTION IN SPOKEN LANGUAGE UNDERSTANDING
    Kuo, Hong-Kwang J.
    Tuske, Zoltan
    Thomas, Samuel
    Kingsbury, Brian
    Saon, George
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7162 - 7166
  • [38] End-to-end latency characterization of task communication models for automotive systems
    Martinez, Jorge
    Sanudo, Ignacio
    Bertogna, Marko
    REAL-TIME SYSTEMS, 2020, 56 (03) : 315 - 347
  • [39] End-to-end latency characterization of task communication models for automotive systems
    Jorge Martinez
    Ignacio Sañudo
    Marko Bertogna
    Real-Time Systems, 2020, 56 : 315 - 347
  • [40] Recent progress in deep end-to-end models for spoken language processing
    Audhkhasi, K.
    Rosenberg, A.
    Saon, G.
    Sethy, A.
    Ramabhadran, B.
    Chen, S.
    Picheny, M.
    IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2017, 61 (4-5)