On Optimal Power Control for URLLC over a Non-stationary Wireless Channel using Contextual Reinforcement Learning

被引:0
|
作者
Sharma, Mohit K.
Sun, Sumei
Kurniawan, Ernest
Tan, Peng Hui
机构
来源
IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022) | 2022年
关键词
Energy minimization; non-stationary wireless channel; reinforcement learning; URLLC; LOW-LATENCY COMMUNICATIONS; COMMUNICATION; OPTIMIZATION; NETWORKS; SYSTEMS;
D O I
10.1109/ICC45855.2022.9839177
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In this work we investigate the design of energy-optimal policies for ultra-reliable low-latency communications (URLLC) over a non-stationary wireless channel, using a contextual reinforcement learning (RL) framework. We consider a point-to-point communication system over a piece-wise stationary wireless channel where the Doppler frequency of the channel switches between two distinct values, depending on the underlying state of the channel. To benchmark the performance, first we consider an oracle agent which has a perfect but causal information about the switching instants, and consists of two deep RL (DRL) agents each of which is tasked with optimal decision making in a unique partially stationary environment. Comparing the performance of the oracle agent with the conventional DRL reveals that the performance gain obtained using oracle agent depends on the dynamics of the non-stationary channel. In particular, for a non-stationary channel with faster switching rate the oracle agent results in approximately 15 - 20% less energy consumption. In contrast, for a channel with slower switching rate the performance of the oracle agent is similar to the conventional DRL agent. Next, for a more realistic scenario when the information about the switching instants for the Doppler frequency of the underlying channel is not available, we model the non-stationary channel as a regime switching process modulated by a Markov process, and adapt the oracle agent by aiding a state tracking algorithm proposed for the regime switching process. Our simulation results show that the proposed algorithm yields a better performance compared to the conventional DRL agent.
引用
收藏
页码:5493 / 5498
页数:6
相关论文
共 50 条
  • [21] Blocklength Allocation and Power Control in UAV-Assisted URLLC System via Multi-agent Deep Reinforcement Learning
    Li, Xinmin
    Zhang, Xuhao
    Li, Jiahui
    Luo, Feiying
    Huang, Yi
    Zhang, Xiaoqiang
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2024, 17 (01)
  • [22] Traffic Scheduling in Non-Stationary Multipath Non-Terrestrial Networks: A Reinforcement Learning Approach
    Machumilane, Achilles
    Gotta, Alberto
    Cassara, Pietro
    Gennaro, Claudio
    Amato, Giuseppe
    ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 4094 - 4099
  • [23] Bandit-Based Power Control in Full-Duplex Cooperative Relay Networks With Strict-Sense Stationary and Non-Stationary Wireless Communication Channels
    Nomikos, Nikolaos
    Talebi, Mohammad Sadegh
    Charalambous, Themistoklis
    Wichman, Risto
    IEEE OPEN JOURNAL OF THE COMMUNICATIONS SOCIETY, 2022, 3 : 366 - 378
  • [24] Optimal Power Control for Over-The-Air Federated Edge Learning Using Statistical Channel Knowledge
    Yu, Xichen
    Xiao, Bingnan
    Ni, Wei
    Wang, Xin
    2022 14TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS AND SIGNAL PROCESSING, WCSP, 2022, : 232 - 237
  • [25] Context Detection and Identification In Multi-Agent Reinforcement Learning With Non-Stationary Environment
    Selamet, Ekrem Talha
    Tumer, Borahan
    2022 30TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU, 2022,
  • [26] Reinforcement learning and evolutionary algorithms for non-stationary multi-armed bandit problems
    Koulouriotis, D. E.
    Xanthopoulos, A.
    APPLIED MATHEMATICS AND COMPUTATION, 2008, 196 (02) : 913 - 922
  • [27] Real-time dynamic pricing in a non-stationary environment using model-free reinforcement learning
    Rana, Rupal
    Oliveira, Fernando S.
    OMEGA-INTERNATIONAL JOURNAL OF MANAGEMENT SCIENCE, 2014, 47 : 116 - 126
  • [28] Reinforcement Learning for Adaptive Optimal Stationary Control of Linear Stochastic Systems
    Pang, Bo
    Jiang, Zhong-Ping
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2023, 68 (04) : 2383 - 2390
  • [29] Power Control for Wireless VBR Video Streaming: From Optimization to Reinforcement Learning
    Ye, Chuang
    Gursoy, M. Cenk
    Velipasalar, Senem
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2019, 67 (08) : 5629 - 5644
  • [30] A General 3-D Non-Stationary 5G Wireless Channel Model
    Wu, Shangbin
    Wang, Cheng-Xiang
    Aggoune, el-Hadi M.
    Alwakeel, Mohammed M.
    You, Xiaohu
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2018, 66 (07) : 3065 - 3078