Graph reinforcement learning with relational priors for predictive power allocation

被引:0
|
作者
Jianyu ZHAO
Chenyang YANG
机构
[1] SchoolofElectronicsandInformationEngineering,BeihangUniversity
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep reinforcement learning for resource allocation has been investigated extensively owing to its ability of handling model-free and end-to-end problems. However, its slow convergence and high time complexity during online training hinder its practical use in dynamic wireless systems. To reduce the training complexity, we resort to graph reinforcement learning for leveraging two kinds of relational priors inherent in many wireless communication problems: topology information and permutation properties. To harness the two priors, we first conceive a method to convert the state matrix into a state graph,and then propose a graph deep deterministic policy gradient(DDPG) algorithm with the desired permutation property. To demonstrate how to apply the proposed methods, we consider a representative problem of using reinforcement learning, predictive power allocation, which minimizes the energy consumption while ensuring the quality-of-service of each user requesting video streaming. We derive the time complexity required by training the proposed graph DDPG algorithm and fully-connected neural network-based DDPG algorithm in each time step. Simulations show that the graph DDPG algorithm converges much faster and needs much lower time and space complexity than existing DDPG algorithms to achieve the same learning performance.
引用
收藏
页码:230 / 247
页数:18
相关论文
共 50 条
  • [1] Graph reinforcement learning with relational priors for predictive power allocation
    Zhao, Jianyu
    Yang, Chenyang
    SCIENCE CHINA-INFORMATION SCIENCES, 2025, 68 (02)
  • [2] Graph kernels and Gaussian processes for relational reinforcement learning
    Gärtner, T
    Driessens, K
    Ramon, J
    INDUCTIVE LOGIC PROGRAMMING, PROCEEDINGS, 2003, 2835 : 146 - 163
  • [3] Graph kernels and Gaussian processes for relational reinforcement learning
    Driessens, Kurt
    Ramon, Jan
    Gaertner, Thomas
    MACHINE LEARNING, 2006, 64 (1-3) : 91 - 119
  • [4] Graph kernels and Gaussian processes for relational reinforcement learning
    Kurt Driessens
    Jan Ramon
    Thomas Gärtner
    Machine Learning, 2006, 64 : 91 - 119
  • [5] Power allocation using spatio-temporal graph neural networks and reinforcement learning
    Jamshidiha, Saeed
    Pourahmadi, Vahid
    Mohammadi, Abbas
    Bennis, Mehdi
    WIRELESS NETWORKS, 2025, 31 (02) : 1163 - 1176
  • [6] Correlation Priors for Reinforcement Learning
    Alt, Bastian
    Sosic, Adrian
    Koeppl, Heinz
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [7] Relational reinforcement learning
    Driessens, K
    AI COMMUNICATIONS, 2005, 18 (01) : 71 - 73
  • [8] Relational reinforcement learning
    Driessens, K
    MULTI-AGENT SYSTEMS AND APPLICATIONS, 2001, 2086 : 271 - 280
  • [9] Relational reinforcement learning
    Dzeroski, S
    De Raedt, L
    Driessens, K
    MACHINE LEARNING, 2001, 43 (1-2) : 7 - 52
  • [10] Relational Reinforcement Learning
    Sašo Džeroski
    Luc De Raedt
    Kurt Driessens
    Machine Learning, 2001, 43 : 7 - 52