Joint Data Transmission and Energy Harvesting for MISO Downlink Transmission Coordination in Wireless IoT Networks

被引:5
作者
Liu, Jain-Shing [1 ]
Lin, Chun-Hung [2 ]
Hu, Yu-Chen [3 ]
Donta, Praveen Kumar [4 ]
机构
[1] Providence Univ, Dept Comp Sci & Informat Engn, Taichung 433, Taiwan
[2] Natl Sun Yat sen Univ, Dept Comp Sci & Engn, Kaohsiung 804, Taiwan
[3] Providence Univ, Dept Comp Sci & Informat Management, Taichung 433, Taiwan
[4] TU Wien, Res Unit Distributed Syst, A-1040 Vienna, Austria
关键词
IoT; SWIPT; joint optimization; beamforming; power control; energy harvesting; transmission coordination; deep reinforcement learning; POWER TRANSFER; INFORMATION; EFFICIENCY; SYSTEM; SWIPT; OPTIMIZATION; MAXIMIZATION; ALLOCATION; DESIGN;
D O I
10.3390/s23083900
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The advent of simultaneous wireless information and power (SWIPT) has been regarded as a promising technique to provide power supplies for an energy sustainable Internet of Things (IoT), which is of paramount importance due to the proliferation of high data communication demands of low-power network devices. In such networks, a multi-antenna base station (BS) in each cell can be utilized to concurrently transmit messages and energies to its intended IoT user equipment (IoT-UE) with a single antenna under a common broadcast frequency band, resulting in a multi-cell multi-input single-output (MISO) interference channel (IC). In this work, we aim to find the trade-off between the spectrum efficiency (SE) and energy harvesting (EH) in SWIPT-enabled networks with MISO ICs. For this, we derive a multi-objective optimization (MOO) formulation to obtain the optimal beamforming pattern (BP) and power splitting ratio (PR), and we propose a fractional programming (FP) model to find the solution. To tackle the nonconvexity of FP, an evolutionary algorithm (EA)-aided quadratic transform technique is proposed, which recasts the nonconvex problem as a sequence of convex problems to be solved iteratively. To further reduce the communication overhead and computational complexity, a distributed multi-agent learning-based approach is proposed that requires only partial observations of the channel state information (CSI). In this approach, each BS is equipped with a double deep Q network (DDQN) to determine the BP and PR for its UE with lower computational complexity based on the observations through a limited information exchange process. Finally, with the simulation experiments, we verify the trade-off between SE and EH, and we demonstrate that, apart from the FP algorithm introduced to provide superior solutions, the proposed DDQN algorithm also shows its performance gain in terms of utility to be up to 1.23-, 1.87-, and 3.45-times larger than the Advantage Actor Critic (A2C), greedy, and random algorithms, respectively, in comparison in the simulated environment.
引用
收藏
页数:24
相关论文
共 53 条
[1]   Antenna Clustering for Simultaneous Wireless Information and Power Transfer in a MIMO Full-Duplex System: A Deep Reinforcement Learning-Based Design [J].
Al-Eryani, Yasser ;
Akrout, Mohamed ;
Hossain, Ekram .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2021, 69 (04) :2331-2345
[2]  
[Anonymous], TR36814 3GPP
[3]  
Bertsekas DP., 2005, DYNAMIC PROGRAMMING
[4]   Multi-Agent Reinforcement Learning: A Review of Challenges and Applications [J].
Canese, Lorenzo ;
Cardarilli, Gian Carlo ;
Di Nunzio, Luca ;
Fazzolari, Rocco ;
Giardino, Daniele ;
Re, Marco ;
Spano, Sergio .
APPLIED SCIENCES-BASEL, 2021, 11 (11)
[5]   Fundamentals of Wireless Information and Power Transfer: From RF Energy Harvester Models to Signal and System Designs [J].
Clerckx, Bruno ;
Zhang, Rui ;
Schober, Robert ;
Ng, Derrick Wing Kwan ;
Kim, Dong In ;
Poor, H. Vincent .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2019, 37 (01) :4-33
[6]   Wireless Information and Power Transfer: Energy Efficiency Optimization in OFDMA Systems [J].
Derrick Wing Kwan Ng ;
Lo, Ernest S. ;
Schober, Robert .
IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2013, 12 (12) :6352-6370
[7]  
Dinkelbach W., 1967, MANAGE SCI, V13, P492
[8]  
Ehrgott M., 2005, MULTICRITERIA OPTIMI
[9]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[10]   Deep Reinforcement Learning for Distributed Dynamic MISO Downlink-Beamforming Coordination [J].
Ge, Jungang ;
Liang, Ying-Chang ;
Joung, Jingon ;
Sun, Sumei .
IEEE TRANSACTIONS ON COMMUNICATIONS, 2020, 68 (10) :6070-6085