Reinforcement Learning Exploration Algorithms for Energy Harvesting Communications Systems

被引:0
|
作者
Masadeh, Ala'eddin [1 ]
Wang, Zhengdao [1 ]
Kamal, Ahmed E. [1 ]
机构
[1] ISU, Ames, IA 50011 USA
来源
2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC) | 2018年
基金
美国国家科学基金会;
关键词
Energy harvesting communications; Markov decision process; Reinforcement learning; Exploration; Exploitation;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Prolonging the lifetime, and maximizing the throughput are important factors in designing an efficient communications system, especially for energy harvesting-based systems. In this work, the problem of maximizing the throughput of point-to-point energy harvesting communications system, while prolonging its lifetime is investigated. This work considers more real communications system, where this system does not have a priori knowledge about the environment. This system consists of a transmitter and receiver. The transmitter is equipped with an infinite buffer to store data, and energy harvesting capability to harvest renewable energy and store it in a finite battery. The problem of finding an efficient power allocation policy is formulated as a reinforcement learning problem. Two different exploration algorithms are used, which are the convergence-based and the epsilon-greedy algorithms. The first algorithm uses the action-value function convergence error and the exploration time threshold to balance between exploration and exploitation. On the other hand, the second algorithm tries to achieve balancing through the exploration probability (i.e. epsilon). Simulation results show that the convergence-based algorithm outperforms the epsilon-greedy algorithm. Then, the effects of the parameters of each algorithm are investigated.
引用
收藏
页数:6
相关论文
共 50 条
  • [41] A Robust Exploration Strategy in Reinforcement Learning Based on Temporal Difference Error
    Hajar, Muhammad Shadi
    Kalutarage, Harsha
    Al-Kadri, M. Omar
    AI 2022: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13728 : 789 - 799
  • [42] Learning to soar: Resource-constrained exploration in reinforcement learning
    Chung, Jen Jen
    Lawrance, Nicholas R. J.
    Sukkarieh, Salah
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2015, 34 (02) : 158 - 172
  • [43] Deep Reinforcement Learning-Based Access Control for Buffer-Aided Relaying Systems With Energy Harvesting
    Zhang, Haodi
    Zhan, Di
    Zhang, Chen Jason
    Wu, Kaishun
    Liu, Ye
    Luo, Sheng
    IEEE ACCESS, 2020, 8 (08): : 145006 - 145017
  • [44] Exploration and Incentives in Reinforcement Learning
    Simchowitz, Max
    Slivkins, Aleksandrs
    OPERATIONS RESEARCH, 2024, 72 (03) : 983 - 998
  • [45] Reinforcement Learning Models and Algorithms for Diabetes Management
    Yau, Kok-Lim Alvin
    Chong, Yung-Wey
    Fan, Xiumei
    Wu, Celimuge
    Saleem, Yasir
    Lim, Phei-Ching
    IEEE ACCESS, 2023, 11 : 28391 - 28415
  • [46] Backtracking Exploration for Reinforcement Learning
    Chen, Xingguo
    Chen, Zening
    Sun, Dingyuanhao
    Gao, Yang
    2023 5TH INTERNATIONAL CONFERENCE ON DISTRIBUTED ARTIFICIAL INTELLIGENCE, DAI 2023, 2023,
  • [47] Reinforcement Learning for Scheduling Wireless Powered Sensor Communications
    Li, Kai
    Ni, Wei
    Abolhasan, Mehran
    Tovar, Eduardo
    IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2019, 3 (02): : 264 - 274
  • [48] REINFORCEMENT LEARNING FOR POWER MANAGEMENT IN WIRELESS MULTIMEDIA COMMUNICATIONS
    Mastronarde, Nicholas
    van der Schaar, Mihaela
    2011 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2011,
  • [49] MODELING THE ENVIRONMENT IN DEEP REINFORCEMENT LEARNING: THE CASE OF ENERGY HARVESTING BASE STATIONS
    Piovesan, Nicola
    Miozzo, Marco
    Dini, Paolo
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 8996 - 9000
  • [50] Routing Selection With Reinforcement Learning for Energy Harvesting Multi-Hop CRN
    He, Xiaoli
    Jiang, Hong
    Song, Yu
    He, Chunlin
    Xiao, He
    IEEE ACCESS, 2019, 7 : 54435 - 54448