Reinforcement Learning Exploration Algorithms for Energy Harvesting Communications Systems

被引:0
|
作者
Masadeh, Ala'eddin [1 ]
Wang, Zhengdao [1 ]
Kamal, Ahmed E. [1 ]
机构
[1] ISU, Ames, IA 50011 USA
来源
2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC) | 2018年
基金
美国国家科学基金会;
关键词
Energy harvesting communications; Markov decision process; Reinforcement learning; Exploration; Exploitation;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Prolonging the lifetime, and maximizing the throughput are important factors in designing an efficient communications system, especially for energy harvesting-based systems. In this work, the problem of maximizing the throughput of point-to-point energy harvesting communications system, while prolonging its lifetime is investigated. This work considers more real communications system, where this system does not have a priori knowledge about the environment. This system consists of a transmitter and receiver. The transmitter is equipped with an infinite buffer to store data, and energy harvesting capability to harvest renewable energy and store it in a finite battery. The problem of finding an efficient power allocation policy is formulated as a reinforcement learning problem. Two different exploration algorithms are used, which are the convergence-based and the epsilon-greedy algorithms. The first algorithm uses the action-value function convergence error and the exploration time threshold to balance between exploration and exploitation. On the other hand, the second algorithm tries to achieve balancing through the exploration probability (i.e. epsilon). Simulation results show that the convergence-based algorithm outperforms the epsilon-greedy algorithm. Then, the effects of the parameters of each algorithm are investigated.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] Deep Reinforcement Learning-Assisted Energy Harvesting Wireless Networks
    Ye, Junliang
    Gharavi, Hamid
    IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2021, 5 (02): : 990 - 1002
  • [32] Wireless Power and Energy Harvesting Control in IoD by Deep Reinforcement Learning
    Yao, Jingjing
    Ansari, Nirwan
    IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2021, 5 (02): : 980 - 989
  • [33] Reinforcement Learning Techniques in Optimizing Energy Systems
    Stavrev, Stefan
    Ginchev, Dimitar
    ELECTRONICS, 2024, 13 (08)
  • [34] Reinforcement Learning with Derivative-Free Exploration
    Chen, Xiong-Hui
    Yu, Yang
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 1880 - 1882
  • [35] Lightweight Reinforcement Learning for Energy Efficient Communications in Wireless Sensor Networks
    Savaglio, Claudio
    Pace, Pasquale
    Aloi, Gianluca
    Liotta, Antonio
    Fortino, Giancarlo
    IEEE ACCESS, 2019, 7 : 29355 - 29364
  • [36] Consensus Algorithms and Deep Reinforcement Learning in Energy Market: A Review
    Jogunola, Olamide
    Adebisi, Bamidele
    Ikpehai, Augustine
    Popoola, Segun, I
    Gui, Guan
    Gacanin, Haris
    Ci, Song
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (06) : 4211 - 4227
  • [37] Exploration With Task Information for Meta Reinforcement Learning
    Jiang, Peng
    Song, Shiji
    Huang, Gao
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4033 - 4046
  • [38] BALANCING EXPLORATION AND EXPLOITATION IN REINFORCEMENT LEARNING USING A VALUE OF INFORMATION CRITERION
    Sledge, Isaac J.
    Principe, Jose C.
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 2816 - 2820
  • [39] Intrinsically Motivated Lifelong Exploration in Reinforcement Learning
    Bougie, Nicolas
    Ichise, Ryutaro
    ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 1357 : 109 - 120
  • [40] Reinforcement Learning for Maritime Communications
    Rong, Bo
    IEEE WIRELESS COMMUNICATIONS, 2023, 30 (03) : 12 - 12