Reinforcement Learning Exploration Algorithms for Energy Harvesting Communications Systems

被引:0
|
作者
Masadeh, Ala'eddin [1 ]
Wang, Zhengdao [1 ]
Kamal, Ahmed E. [1 ]
机构
[1] ISU, Ames, IA 50011 USA
来源
2018 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC) | 2018年
基金
美国国家科学基金会;
关键词
Energy harvesting communications; Markov decision process; Reinforcement learning; Exploration; Exploitation;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Prolonging the lifetime, and maximizing the throughput are important factors in designing an efficient communications system, especially for energy harvesting-based systems. In this work, the problem of maximizing the throughput of point-to-point energy harvesting communications system, while prolonging its lifetime is investigated. This work considers more real communications system, where this system does not have a priori knowledge about the environment. This system consists of a transmitter and receiver. The transmitter is equipped with an infinite buffer to store data, and energy harvesting capability to harvest renewable energy and store it in a finite battery. The problem of finding an efficient power allocation policy is formulated as a reinforcement learning problem. Two different exploration algorithms are used, which are the convergence-based and the epsilon-greedy algorithms. The first algorithm uses the action-value function convergence error and the exploration time threshold to balance between exploration and exploitation. On the other hand, the second algorithm tries to achieve balancing through the exploration probability (i.e. epsilon). Simulation results show that the convergence-based algorithm outperforms the epsilon-greedy algorithm. Then, the effects of the parameters of each algorithm are investigated.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] Online Learning of Power Allocation Policies in Energy Harvesting Communications
    Sakulkar, Pranav
    Krishnamachari, Bhaskar
    2016 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS (SPCOM), 2016,
  • [22] Distributed Reinforcement Learning Algorithm for Energy Harvesting Sensor Networks
    Al-Tous, Hanan
    Barhumi, Imad
    2019 IEEE INTERNATIONAL BLACK SEA CONFERENCE ON COMMUNICATIONS AND NETWORKING (BLACKSEACOM), 2019,
  • [23] Energy-Harvesting Strategy Investigation for Glider Autonomous Soaring Using Reinforcement Learning
    Zhao, Jiachi
    Li, Jun
    Zeng, Lifang
    AEROSPACE, 2023, 10 (10)
  • [24] Improving Reinforcement Learning Exploration by Autoencoders
    Paczolay, Gabor
    Harmati, Istvan
    Periodica Polytechnica Electrical Engineering and Computer Science, 2024, 68 (04): : 335 - 343
  • [25] On the Design of Tailored Neural Networks for Energy Harvesting Broadcast Channels: A Reinforcement Learning Approach
    Kim, Heasung
    Kim, Jungtai
    Shin, Wonjae
    Yang, Heecheol
    Lee, Nayoung
    Kim, Seong Jin
    Lee, Jungwoo
    IEEE ACCESS, 2020, 8 : 179678 - 179691
  • [26] Optimal Transmission Policy in Energy Harvesting Wireless Communications: A Learning Approach
    Wu, Keyu
    Tellambura, Chintha
    Jiang, Hai
    2017 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2017,
  • [27] Exploration-Driven Genetic Algorithms for Hyperparameter Optimisation in Deep Reinforcement Learning
    Brzek, Bartlomiej
    Probierz, Barbara
    Kozak, Jan
    APPLIED SCIENCES-BASEL, 2025, 15 (04):
  • [28] POLICY AUGMENTATION: AN EXPLORATION STRATEGY FOR FASTER CONVERGENCE OF DEEP REINFORCEMENT LEARNING ALGORITHMS
    Mahyari, Arash
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3505 - 3509
  • [29] Power Allocation Based on Reinforcement Learning for MIMO System With Energy Harvesting
    Mu, Xingchi
    Zhao, Xiaohui
    Liang, Hui
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (07) : 7622 - 7633
  • [30] Power Control in Energy Harvesting Multiple Access System with Reinforcement Learning
    Chu, Man
    Liao, Xuewen
    Li, Hang
    Cui, Shuguang
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,