Stable Reinforcement Learning for Optimal Frequency Control: A Distributed Averaging-Based Integral Approach

被引:8
作者
Jiang, Yan [1 ]
Cui, Wenqi [1 ]
Zhang, Baosen [1 ]
Cortes, Jorge [2 ]
机构
[1] Univ Washington, Dept Elect & Comp Engn, Seattle, WA 98195 USA
[2] Univ Calif San Diego, Dept Mech & Aerosp Engn, La Jolla, CA 92093 USA
来源
IEEE OPEN JOURNAL OF CONTROL SYSTEMS | 2022年 / 1卷
关键词
Frequency control; Lyapunov stability; reinforcement learning; steady-state and transient performance; ECONOMIC-DISPATCH; POWER-SYSTEM; STABILITY;
D O I
10.1109/OJCSYS.2022.3202202
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Frequency control plays a pivotal role in reliable power system operations. It is conventionally performed in a hierarchical way that first rapidly stabilizes the frequency deviations and then slowly recovers the nominal frequency. However, as the generation mix shifts from synchronous generators to renewable resources, power systems experience larger and faster frequency fluctuations due to the loss of inertia, which adversely impacts the frequency stability. This has motivated active research in algorithms that jointly address frequency degradation and economic efficiency in a fast timescale, among which the distributed averaging-based integral (DAI) control is a notable one that sets controllable power injections directly proportional to the integrals of frequency deviation and economic inefficiency signals. Nevertheless, DAI does not typically consider the transient performance of the system following power disturbances and has been restricted to quadratic operational cost functions. This paper aims to leverage nonlinear optimal controllers to simultaneously achieve optimal transient frequency control and find the most economic power dispatch for frequency restoration. To this end, we integrate reinforcement learning (RL) to the classic DAI, which results in RL-DAI control. Specifically, we use RL to learn a neural network-based control policy mapping from the integral variables of DAI to the controllable power injections which provides optimal transient frequency control, while DAI inherently ensures the frequency restoration and optimal economic dispatch. Compared to existing methods, we provide provable guarantees on the stability of the learned controllers and extend the set of allowable cost functions to a much larger class. Simulations on the 39-bus New England system illustrate our results.
引用
收藏
页码:194 / 209
页数:16
相关论文
共 39 条
[1]   PRACTICAL METHOD FOR THE DIRECT ANALYSIS OF TRANSIENT STABILITY [J].
ATHAY, T ;
PODMORE, R ;
VIRMANI, S .
IEEE TRANSACTIONS ON POWER APPARATUS AND SYSTEMS, 1979, 98 (02) :573-584
[2]  
Binmore K., 1977, Mathematical analysis: a straightforward approach
[3]  
Bird L., 2013, Integrating variable renewable energy: Challenges and solutions
[4]  
Boyd S., 2004, CONVEX OPTIMIZATION
[5]  
Bullo F., 2022, LECT NETWORK SYSTEMS
[6]   Model-Free Emergency Frequency Control Based on Reinforcement Learning [J].
Chen, Chunyu ;
Cui, Mingjian ;
Li, Fangxing ;
Yin, Shengfei ;
Wang, Xinan .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (04) :2336-2346
[7]   Reinforcement Learning for Selective Key Applications in Power Systems: Recent Advances and Future Challenges [J].
Chen, Xin ;
Qu, Guannan ;
Tang, Yujie ;
Low, Steven ;
Li, Na .
IEEE TRANSACTIONS ON SMART GRID, 2022, 13 (04) :2935-2958
[8]   Frequency-driven market mechanisms for optimal dispatch in power networks [J].
Cherukuri, Ashish ;
Stegink, Tjerk ;
De Persis, Claudio ;
van der Schaft, Arjan ;
Cortes, Jorge .
AUTOMATICA, 2021, 133
[9]   A TOOLBOX FOR POWER-SYSTEM DYNAMICS AND CONTROL ENGINEERING-EDUCATION AND RESEARCH [J].
CHOW, JH ;
CHEUNG, KW .
IEEE TRANSACTIONS ON POWER SYSTEMS, 1992, 7 (04) :1559-1564
[10]   Reinforcement Learning for Optimal Primary Frequency Control: A Lyapunov Approach [J].
Cui, Wenqi ;
Jiang, Yan ;
Zhang, Baosen .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2023, 38 (02) :1676-1688