Knowledge Distillation-Based Semantic Communications for Multiple Users

被引:2
作者
Liu, Chenguang [1 ]
Zhou, Yuxin [2 ]
Chen, Yunfei [3 ]
Yang, Shuang-Hua [4 ,5 ]
机构
[1] Univ Warwick, Sch Engn, Coventry CV4 7AL, England
[2] Southern Univ Sci & Technol, Dept Comp Sci & Engn, Shenzhen 518055, Peoples R China
[3] Univ Durham, Dept Engn, Durham DH1 3LE, England
[4] Southern Univ Sci & Technol, Shenzhen Key Lab Safety & Secur Next Generat Ind I, Shenzhen 518055, Peoples R China
[5] Univ Reading, Dept Comp Sci, Reading RG6 6AB, England
基金
中国国家自然科学基金; 欧盟地平线“2020”;
关键词
Interference; Semantics; Complexity theory; Data models; Communication systems; Training; Task analysis; Deep learning; knowledge distillation; model compression; multi-user interference; semantic communication; text transmission; INTERNET;
D O I
10.1109/TWC.2023.3336941
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep learning (DL) has shown great potential in revolutionizing the traditional communications system. Many applications in communications have adopted DL techniques due to their powerful representation ability. However, the learning-based methods can be dependent on the training dataset and perform worse on unseen interference due to limited model generalizability and complexity. In this paper, we consider the semantic communication (SemCom) system with multiple users, where there is a limited number of training samples and unexpected interference. To improve the model generalization ability and reduce the model size, we propose a knowledge distillation (KD) based system where Transformer based encoder-decoder is implemented as the semantic encoder-decoder and fully connected neural networks are implemented as the channel encoder-decoder. Specifically, four types of knowledge transfer and model compression are analyzed. Important system and model parameters are considered, including the level of noise and interference, the number of interfering users and the size of the encoder and decoder. Numerical results demonstrate that KD significantly improves the robustness and the generalization ability when applied to unexpected interference, and it reduces the performance loss when compressing the model size.
引用
收藏
页码:7000 / 7012
页数:13
相关论文
共 34 条
[11]  
Jiao XQ, 2020, FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, P4163
[12]  
Kingma Diederik P, 2014, ARXIV PREPRINT ARXIV
[13]   ON INFORMATION AND SUFFICIENCY [J].
KULLBACK, S ;
LEIBLER, RA .
ANNALS OF MATHEMATICAL STATISTICS, 1951, 22 (01) :79-86
[14]   DeepJS']JSCC-f : Deep Joint Source-Channel Coding of Images With Feedback [J].
Kurka, David Burth ;
Gunduz, Deniz .
IEEE JOURNAL ON SELECTED AREAS IN INFORMATION THEORY, 2020, 1 (01) :178-193
[15]   Deep Learning Based Detection for Communications Systems With Radar Interference [J].
Liu, Chenguang ;
Chen, Yunfei ;
Yang, Shuang-Hua .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (06) :6245-6254
[16]   Signal detection with co-channel interference using deep learning [J].
Liu, Chenguang ;
Chen, Yunfei ;
Yang, Shuang-Hua .
PHYSICAL COMMUNICATION, 2021, 47
[17]   An Introduction to Deep Learning for the Physical Layer [J].
O'Shea, Timothy ;
Hoydis, Jakob .
IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2017, 3 (04) :563-575
[18]  
Omri A, 2013, IEEE ICC, P5651, DOI 10.1109/ICC.2013.6655494
[19]   BLEU: a method for automatic evaluation of machine translation [J].
Papineni, K ;
Roukos, S ;
Ward, T ;
Zhu, WJ .
40TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, PROCEEDINGS OF THE CONFERENCE, 2002, :311-318
[20]  
Qiao Lan, 2021, Journal of Communications and Information Networks, V6, P336, DOI 10.23919/JCIN.2021.9663101