Distributed Attention-Enabled Multi-Agent Reinforcement Learning Based Frequency Regulation of Power Systems

被引:0
作者
Zhao, Yunzheng [1 ]
Liu, Tao [1 ]
Hill, David J. [1 ,2 ]
机构
[1] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Peoples R China
[2] Monash Univ, Dept Elect & Comp Syst Engn, Clayton, Vic 3800, Australia
关键词
Frequency control; Training; Power system stability; Generators; Reinforcement learning; Time-varying systems; Renewable energy sources; Neural networks; Topology; Scalability; Distributed attention-enabled multi-agent reinforcement learning; frequency regulation; AUTOMATIC-GENERATION CONTROL;
D O I
10.1109/TPWRS.2024.3469132
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This paper develops a new distributed attention-enabled multi-agent reinforcement learning method for frequency regulation of power systems. Specifically, the controller of each generator is modelled as an agent, and the reward and observation are designed based on the characteristics of power systems. All the agents learn their own control policies in the offline training phase and generate frequency control signals in the online execution phase. The target of the proposed algorithm is to conduct both offline training and online frequency control in a distributed way. To achieve this goal, two distributed information-sharing mechanisms are proposed based on the different global information to be discovered. First, a consensus-based reward-sharing mechanism is designed to estimate the globally averaged reward. Second, a distributed observation-sharing scheme is developed to discover the global observation information. Furthermore, the attention strategy is embedded in the observation-sharing scheme to help agents adaptively adjust the importance of observations from different neighbors. With these two mechanisms, a new distributed attention-enabled proximal policy optimization (DAPPO) based method is proposed to achieve model-free frequency control. Simulation results on the IEEE 39-bus system and the NPCC 140-bus system demonstrate that the proposed DAPPO achieves stable offline training and effective online frequency control.
引用
收藏
页码:2427 / 2437
页数:11
相关论文
共 35 条
[1]  
Andreasson M, 2014, P AMER CONTR CONF, P3183, DOI 10.1109/ACC.2014.6858999
[2]   Distributed Control of Networked Dynamical Systems: Static Feedback, Integral Action and Consensus [J].
Andreasson, Martin ;
Dimarogonas, Dimos V. ;
Sandberg, Henrik ;
Johansson, Karl Henrik .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2014, 59 (07) :1750-1764
[3]   PRACTICAL METHOD FOR THE DIRECT ANALYSIS OF TRANSIENT STABILITY [J].
ATHAY, T ;
PODMORE, R ;
VIRMANI, S .
IEEE TRANSACTIONS ON POWER APPARATUS AND SYSTEMS, 1979, 98 (02) :573-584
[4]  
Australian Energy Market Commission (AEMC), 2023, Electricity guidelines, standards and schedules
[5]  
Bahdanau D, 2016, Arxiv, DOI arXiv:1409.0473
[6]  
Bevrani H, 2014, POWER ELECTRON POWER, P1, DOI 10.1007/978-3-319-07278-4
[7]   Attention Enabled Multi-Agent DRL for Decentralized Volt-VAR Control of Active Distribution System Using PV Inverters and SVCs [J].
Cao, Di ;
Zhao, Junbo ;
Hu, Weihao ;
Ding, Fei ;
Huang, Qi ;
Chen, Zhe .
IEEE TRANSACTIONS ON SUSTAINABLE ENERGY, 2021, 12 (03) :1582-1592
[8]   Diffusion recursive least-squares for distributed estimation over adaptive networks [J].
Cattivelli, Federico S. ;
Lopes, Cassio G. ;
Sayed, Ali. H. .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2008, 56 (05) :1865-1877
[9]   PowerNet: Multi-Agent Deep Reinforcement Learning for Scalable Powergrid Control [J].
Chen, Dong ;
Chen, Kaian ;
Li, Zhaojian ;
Chu, Tianshu ;
Yao, Rui ;
Qiu, Feng ;
Lin, Kaixiang .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2022, 37 (02) :1007-1017
[10]  
Chu T., 2020, P INT C LEARN REPR