In-Context Learning for MIMO Equalization Using Transformer-Based Sequence Models

被引:0
|
作者
Zecchin, Matteo [1 ]
Yu, Kai [2 ]
Simeone, Osvaldo [1 ]
机构
[1] Kings Coll London, Dept Engn, Ctr Intelligent Informat Proc Syst CIIPS, Kings Commun Learning & Informat Proc KCLIP Lab, London WC2R 2LS, England
[2] Nanjing Univ, Sch Elect Sci & Engn, Nanjing, Peoples R China
来源
2024 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS, ICC WORKSHOPS 2024 | 2024年
基金
英国工程与自然科学研究理事会;
关键词
Machine learning; wireless communications; meta-learning; large language models; transformer; in-context learning;
D O I
10.1109/ICCWORKSHOPS59551.2024.10615360
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Large pre-trained sequence models, such as transformer-based architectures, have been recently shown to have the capacity to carry out in-context learning (ICL). In ICL, a decision on a new input is made via a direct mapping of the input and of a few examples from the given task, serving as the task's context, to the output variable. No explicit updates of the model parameters are needed to tailor the decision to a new task. Pre-training, which amounts to a form of meta-learning, is based on the observation of examples from several related tasks. Prior work has shown ICL capabilities for linear regression. In this study, we leverage ICL to address the inverse problem of multiple-input and multiple-output (MIMO) equalization based on a context given by pilot symbols. A task is defined by the unknown fading channel and by the signal-to-noise ratio (SNR) level, which may be known. To highlight the practical potential of the approach, we allow the presence of quantization of the received signals. We demonstrate via numerical results that transformer-based ICL has a threshold behavior, whereby, as the number of pre-training tasks grows, the performance switches from that of a minimum mean squared error (MMSE) equalizer with a prior determined by the pre-trained tasks to that of an MMSE equalizer with the true data-generating prior.
引用
收藏
页码:1573 / 1578
页数:6
相关论文
共 50 条
  • [21] Automatic text summarization using transformer-based language models
    Rao, Ritika
    Sharma, Sourabh
    Malik, Nitin
    INTERNATIONAL JOURNAL OF SYSTEM ASSURANCE ENGINEERING AND MANAGEMENT, 2024, 15 (06) : 2599 - 2605
  • [22] Accelerating Transformer-based Deep Learning Models on FPGAs using Column Balanced Block Pruning
    Peng, Hongwu
    Huang, Shaoyi
    Geng, Tong
    Li, Ang
    Jiang, Weiwen
    Liu, Hang
    Wang, Shusen
    Ding, Caiwen
    PROCEEDINGS OF THE 2021 TWENTY SECOND INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN (ISQED 2021), 2021, : 142 - 148
  • [23] Precipitation nowcasting using transformer-based generative models and transfer learning for improved disaster preparedness
    Piran, Md. Jalil
    Wang, Xiaoding
    Kim, Ho Jun
    Kwon, Hyun Han
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2024, 132
  • [24] Promises and perils of using Transformer-based models for SE research
    Xiao, Yan
    Zuo, Xinyue
    Lu, Xiaoyue
    Dong, Jin Song
    Cao, Xiaochun
    Beschastnikh, Ivan
    NEURAL NETWORKS, 2025, 184
  • [25] Adaptive In-Context Learning with Large Language Models for Bundle
    Sun, Zhu
    Feng, Kaidong
    Yang, Jie
    Qu, Xinghua
    Fang, Hui
    Ong, Yew-Soon
    Liu, Wenyuan
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 966 - 976
  • [26] Learning to Retrieve In-Context Examples for Large Language Models
    Wang, Liang
    Yang, Nan
    Wei, Furu
    PROCEEDINGS OF THE 18TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 1752 - 1767
  • [27] Using In-Context Learning to Improve Dialogue Safety
    Meade, Nicholas
    Gella, Spandana
    Hazarika, Devamanyu
    Gupta, Prakhar
    Jin, Di
    Reddy, Siva
    Liu, Yang
    Hakkani-Tur, Dilek
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 11882 - 11910
  • [28] Abusive Content Detection in Arabic Tweets Using Multi-Task Learning and Transformer-Based Models
    Alrashidi, Bedour
    Jamal, Amani
    Alkhathlan, Ali
    APPLIED SCIENCES-BASEL, 2023, 13 (10):
  • [29] Transfer Learning of Transformer-Based Speech Recognition Models from Czech to Slovak
    Lehecka, Jan
    Psutka, Josef, V
    Psutka, Josef
    TEXT, SPEECH, AND DIALOGUE, TSD 2023, 2023, 14102 : 328 - 338
  • [30] A performance analysis of transformer-based deep learning models for Arabic image captioning
    Alsayed, Ashwaq
    Qadah, Thamir M.
    Arif, Muhammad
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (09)