On the Uses of Large Language Models to Design End-to-end Learning Semantic Communication

被引:1
作者
Wang, Ying [1 ]
Sun, Zhuo [1 ]
Fan, Jinpo [1 ]
Ma, Hao [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Key Lab Univ Wireless Commun, Minist Educ, Beijing 100876, Peoples R China
来源
2024 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE, WCNC 2024 | 2024年
基金
中国国家自然科学基金;
关键词
end-to-end learning; large language models; semantic communication;
D O I
10.1109/WCNC57260.2024.10570717
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning-based semantic communication is a promising research direction for next-generation communication systems. The emergence of large language models(LLMs) with remarkable semantic comprehension abilities leads us to consider whether LLMs can be used in semantic communication to enhance model's performance. In this paper, we discuss the main implementing details of the idea by proposing a general end-to-end learning semantic communication model with LLM, including subword-level tokenization, a rate adapter based on gradients for matching the rate requirements of any channel codec and fine-tuning for possessing private background knowledge. By taking Bidirectional and Auto-Regressive Transformers (BART) and Generative Pre-trained Transformer 2 (GPT2) as examples, we demonstrate how we can utilize various structures of LLMs to design semantic codecs. In terms of semantic fidelity, generalizability to cross-scenario, and complexity, results reveal that the LLM-based semantic communication system achieves exciting performance. We hope this initial work can inspire more research devoted to this field.
引用
收藏
页数:6
相关论文
共 14 条
[1]  
Europarl Koehn P., 2005, P MACH TRANSL SUMM 1, P79
[2]  
Francis W. N., 1979, Letters to the Editor, V5, P7
[3]  
Guo S., 2023, IEEE COMMUNICATIONS
[4]  
Jin Chen, 2023, ARXIV
[5]  
Lee J.Y., 2022, ARXIV
[6]  
Lewis M., 2019, arXiv
[7]   BLEU: a method for automatic evaluation of machine translation [J].
Papineni, K ;
Roukos, S ;
Ward, T ;
Zhu, WJ .
40TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, 2002, :311-318
[8]   A Robust Deep Learning Enabled Semantic Communication System for Text [J].
Peng, Xiang ;
Qin, Zhijin ;
Huang, Danlan ;
Tao, Xiaoming ;
Lu, Jianhua ;
Liu, Guangyi ;
Pan, Chengkang .
2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, :2704-2709
[9]  
Radford A, 2019, OPENAI BLOG
[10]  
Selvaraju RR, 2020, INT J COMPUT VISION, V128, P336, DOI [10.1109/ICCV.2017.74, 10.1007/s11263-019-01228-7]