Graph reinforcement learning with relational priors for predictive power allocation

被引:0
|
作者
Jianyu ZHAO
Chenyang YANG
机构
[1] SchoolofElectronicsandInformationEngineering,BeihangUniversity
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning for resource allocation has been investigated extensively owing to its ability of handling model-free and end-to-end problems. However, its slow convergence and high time complexity during online training hinder its practical use in dynamic wireless systems. To reduce the training complexity, we resort to graph reinforcement learning for leveraging two kinds of relational priors inherent in many wireless communication problems: topology information and permutation properties. To harness the two priors, we first conceive a method to convert the state matrix into a state graph,and then propose a graph deep deterministic policy gradient(DDPG) algorithm with the desired permutation property. To demonstrate how to apply the proposed methods, we consider a representative problem of using reinforcement learning, predictive power allocation, which minimizes the energy consumption while ensuring the quality-of-service of each user requesting video streaming. We derive the time complexity required by training the proposed graph DDPG algorithm and fully-connected neural network-based DDPG algorithm in each time step. Simulations show that the graph DDPG algorithm converges much faster and needs much lower time and space complexity than existing DDPG algorithms to achieve the same learning performance.
引用
收藏
页码:230 / 247
页数:18
相关论文
共 50 条
  • [41] Utile Distinctions for Relational Reinforcement Learning
    Dabney, William
    McGovern, Amy
    20TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2007, : 738 - 743
  • [42] Integrating guidance into relational reinforcement learning
    Driessens, K
    Dzeroski, S
    MACHINE LEARNING, 2004, 57 (03) : 271 - 304
  • [43] Relational Reinforcement Learning in Infinite Mario
    Mohan, Shiwali
    Laird, John E.
    PROCEEDINGS OF THE TWENTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE (AAAI-10), 2010, : 1953 - 1954
  • [44] Integrating Guidance into Relational Reinforcement Learning
    Kurt Driessens
    Sašo Džeroski
    Machine Learning, 2004, 57 : 271 - 304
  • [45] Relational Verification using Reinforcement Learning
    Chen, Jia
    Wei, Jiayi
    Feng, Yu
    Bastani, Osbert
    Dillig, Isil
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2019, 3 (OOPSLA):
  • [46] Relational macros for transfer in reinforcement learning
    Torrey, Lisa
    Shavlik, Jude
    Walker, Trevor
    Maclin, Richard
    INDUCTIVE LOGIC PROGRAMMING, 2008, 4894 : 254 - +
  • [47] Inverse Reinforcement Learning in Relational Domains
    Munzer, Thibaut
    Piot, Bilal
    Geist, Matthieu
    Pietquin, Olivier
    Lopes, Manuel
    PROCEEDINGS OF THE TWENTY-FOURTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI), 2015, : 3735 - 3741
  • [48] Safe Reinforcement Learning for Power Allocation of Hybrid Energy Storage Systems in Electric Vehicles Using PPO and Predictive Safety Filter
    Xu, Xuetong
    Zhang, Lijun
    Lai, Jingang
    Yang, Longzhi
    Xiao, Jiangwen
    IEEE TRANSACTIONS ON TRANSPORTATION ELECTRIFICATION, 2025, 11 (01): : 4024 - 4034
  • [49] POWER ALLOCATION FOR WIRELESS FEDERATED LEARNING USING GRAPH NEURAL NETWORKS
    Li, Boning
    Swami, Ananthram
    Segarra, Santiago
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 5243 - 5247
  • [50] Graph Representations for Reinforcement Learning
    Schab, Esteban
    Casanova, Carlos
    Piccoli, Fabiana
    JOURNAL OF COMPUTER SCIENCE & TECHNOLOGY, 2024, 24 (01): : 29 - 38