In-Context Learning for MIMO Equalization Using Transformer-Based Sequence Models

被引:0
|
作者
Zecchin, Matteo [1 ]
Yu, Kai [2 ]
Simeone, Osvaldo [1 ]
机构
[1] Kings Coll London, Dept Engn, Ctr Intelligent Informat Proc Syst CIIPS, Kings Commun Learning & Informat Proc KCLIP Lab, London WC2R 2LS, England
[2] Nanjing Univ, Sch Elect Sci & Engn, Nanjing, Peoples R China
来源
2024 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS, ICC WORKSHOPS 2024 | 2024年
基金
英国工程与自然科学研究理事会;
关键词
Machine learning; wireless communications; meta-learning; large language models; transformer; in-context learning;
D O I
10.1109/ICCWORKSHOPS59551.2024.10615360
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Large pre-trained sequence models, such as transformer-based architectures, have been recently shown to have the capacity to carry out in-context learning (ICL). In ICL, a decision on a new input is made via a direct mapping of the input and of a few examples from the given task, serving as the task's context, to the output variable. No explicit updates of the model parameters are needed to tailor the decision to a new task. Pre-training, which amounts to a form of meta-learning, is based on the observation of examples from several related tasks. Prior work has shown ICL capabilities for linear regression. In this study, we leverage ICL to address the inverse problem of multiple-input and multiple-output (MIMO) equalization based on a context given by pilot symbols. A task is defined by the unknown fading channel and by the signal-to-noise ratio (SNR) level, which may be known. To highlight the practical potential of the approach, we allow the presence of quantization of the received signals. We demonstrate via numerical results that transformer-based ICL has a threshold behavior, whereby, as the number of pre-training tasks grows, the performance switches from that of a minimum mean squared error (MMSE) equalizer with a prior determined by the pre-trained tasks to that of an MMSE equalizer with the true data-generating prior.
引用
收藏
页码:1573 / 1578
页数:6
相关论文
共 50 条
  • [1] In-Context In-Context Learning with Transformer Neural Processes
    Ashman, Matthew
    Diaconu, Cristiana
    Weller, Adrian
    Turner, Richard E.
    SYMPOSIUM ON ADVANCES IN APPROXIMATE BAYESIAN INFERENCE, 2024, 253 : 1 - 29
  • [2] Transformer-based in-context policy learning for efficient active flow control across various airfoils
    Zheng, Changdong
    Xie, Fangfang
    Ji, Tingwei
    Zhou, Hongjie
    Zheng, Yao
    JOURNAL OF FLUID MECHANICS, 2024, 1001
  • [3] Detecting Sarcasm in Conversation Context Using Transformer-Based Models
    Avvaru, Adithya
    Vobilisetty, Sanath
    Mamidi, Radhika
    FIGURATIVE LANGUAGE PROCESSING, 2020, : 98 - 103
  • [4] In-Context Learning Unlocked for Diffusion Models
    Wang, Zhendong
    Jiang, Yifan
    Lu, Yadong
    Shen, Yelong
    He, Pengcheng
    Chen, Weizhu
    Wang, Zhangyang
    Zhou, Mingyuan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [5] Transformer-Based Federated Learning Models for Recommendation Systems
    Reddy, M. Sujaykumar
    Karnati, Hemanth
    Sundari, L. Mohana
    IEEE ACCESS, 2024, 12 : 109596 - 109607
  • [6] Automatic summarization of cooking videos using transfer learning and transformer-based models
    P. M. Alen Sadique
    R. V. Aswiga
    Discover Artificial Intelligence, 5 (1):
  • [7] The Shape of Learning: Anisotropy and Intrinsic Dimensions in Transformer-Based Models
    Razzhigaev, Anton
    Mikhalchuk, Matvey
    Goncharova, Elizaveta
    Oseledets, Ivan
    Dimitrov, Denis
    Kuznetsov, Andrey
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: EACL 2024, 2024, : 868 - 874
  • [8] Classification of Highly Divergent Viruses from DNA/RNA Sequence Using Transformer-Based Models
    Sadad, Tariq
    Aurangzeb, Raja Atif
    Imran
    Safran, Mejdl
    Alfarhood, Sultan
    Kim, Jungsuk
    BIOMEDICINES, 2023, 11 (05)
  • [9] Abusive Bangla comments detection on Facebook using transformer-based deep learning models
    Tanjim Taharat Aurpa
    Rifat Sadik
    Md Shoaib Ahmed
    Social Network Analysis and Mining, 2022, 12
  • [10] Dementia Detection using Transformer-Based Deep Learning and Natural Language Processing Models
    Saltz, Ploypaphat
    Lin, Shih Yin
    Cheng, Sunny Chieh
    Si, Dong
    2021 IEEE 9TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2021), 2021, : 509 - 510