Wireless Power Control via Meta-Reinforcement Learning

被引:3
作者
Lu, Ziyang [1 ]
Gursoy, M. Cenk [1 ]
机构
[1] Syracuse Univ, Dept Elect Engn & Comp Sci, Syracuse, NY 13244 USA
来源
IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022) | 2022年
关键词
meta-reinforcement learning; model-agnostic meta-learning; power control; wireless interference networks;
D O I
10.1109/ICC45855.2022.9839179
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
In this paper, the power control problem is addressed in a wireless interference network in which there exist multiple transmitter-receiver pairs sharing the same bandwidth for information exchange. The goal is to train a common deep neural network (DNN) for power allocation at each transmitter. Recent studies in the literature have addressed this problem via deep reinforcement learning (DRL). However, training DRL algorithms can become costly in wireless networks since the DRL algorithm may converge slowly in specific problems and hence require a large amount of training data. Besides, the converged model may fail in a new environment, which is not preferable in a wireless network due to its dynamic and time-varying nature. In this work, we address these considerations by proposing a meta-DRL framework that incorporates the method of Model-Agnostic Meta-Learning (MAML). Within the proposed framework, a common initialization is trained for similar power control tasks. From the initialization, we show that only a few gradient descent steps are required for adapting to an unseen task. Simulation results demonstrate that the proposed framework can outperform conventional DRL and joint-learning (which trains a global model for similar tasks) for power control in wireless interference networks.
引用
收藏
页码:1562 / 1567
页数:6
相关论文
共 12 条
[1]  
[Anonymous], 2017, ARXIV170303400
[2]  
Antoniou A., 2018, ICLR
[3]  
Chen W.Y., 2019, INT C LEARNING REPRE
[4]  
Fallah A, 2020, PR MACH LEARN RES, V108, P1082
[5]  
Fletscher L. A., 2018, IEEE T NETW SERV MAN, V16, P279
[6]   Centralized and Distributed Deep Reinforcement Learning Methods for Downlink Sum-Rate Optimization [J].
Khan, Ahmad Ali ;
Adve, Raviraj S. .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2020, 19 (12) :8410-8426
[7]   The Status of Occupational Burnout and Its Influence on the Psychological Health of Factory Workers and Miners in Wulumuqi, China [J].
Lu, Yaoqin ;
Zhang, Zhe ;
Gao, Sunyujie ;
Yan, Huan ;
Zhang, Lijiang ;
Liu, Jiwen .
BIOMED RESEARCH INTERNATIONAL, 2020, 2020
[8]  
Mnih V., 2013, CoRR abs/1312.5602
[9]   Multi-Agent Deep Reinforcement Learning for Dynamic Power Allocation in Wireless Networks [J].
Nasir, Yasar Sinan ;
Guo, Dongning .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2019, 37 (10) :2239-2250
[10]  
Nichol Alex, Reptile: A Scalable Meta-learning Algorithm