Off-policy deep reinforcement learning with automatic entropy adjustment for adaptive online grid emergency control

被引:2
作者
Zhang, Ying [1 ]
Yue, Meng [1 ]
Wang, Jianhui [2 ]
机构
[1] Brookhaven Natl Lab, Interdisciplinary Sci Dept, Upton, NY 11973 USA
[2] Southern Methodist Univ, Dept Elect & Comp Engn, Dallas, TX 75205 USA
关键词
Deep reinforcement learning; Grid emergency control; Soft actor -critic; Voltage stability; VOLTAGE CONTROL; POWER; DECISION;
D O I
10.1016/j.epsr.2023.109136
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Electric overloading conditions and contingencies put modern power systems at risk of voltage collapse and blackouts. Load shedding is crucial to maintain voltage stability for grid emergency control. However, the rule -or model-based schemes rely on accurate dynamic system models and face considerable challenges in adapting to various operating conditions and uncertain event occurrences. To address these issues, this paper proposes a novel deep reinforcement learning (DRL)-based voltage stability control algorithm with automatic entropy adjustment (AEA) for grid emergency control. Various dynamic network components for complex system op-erations are modeled to construct the DRL environment. An off-policy soft actor-critic architecture is developed to maximize the expected reward and policy entropy simultaneously. The AEA mechanism is proposed to facilitate the policy maximum entropy procedure, and the proposed method can automatically provide effective discrete and continuous actions against various fault scenarios. Our approach accomplishes high sampling ef-ficiency, scalability, and auto-adaptivity of the control policies under high uncertainties. Comparative studies with the existing DRL-based control methods in IEEE benchmarks indicate salient performance improvement of the proposed method for dynamic system emergency control.
引用
收藏
页数:10
相关论文
共 31 条
  • [1] [Anonymous], 2009, Convex Optimization
  • [2] [Anonymous], 2022, RL DUAL GRADIENT DES
  • [3] Bevrani H, 2014, POWER SYSTEM MONITORING AND CONTROL, P1, DOI 10.1002/9781118852422
  • [4] Chow J. H., Power system toolbox (PST)
  • [5] Dalal G, 2018, Arxiv, DOI arXiv:1801.08757
  • [6] Electric Reliability Council of Texas (ERCOT), 2021, Extreme Cold Weather Event: Preliminary Report on Causes of Genera�tor Outages and Derates'
  • [7] Power systems stability control: Reinforcement learning framework
    Ernst, D
    Glavic, M
    Wehenkel, L
    [J]. IEEE TRANSACTIONS ON POWER SYSTEMS, 2004, 19 (01) : 427 - 435
  • [8] Decision Tree-Based Preventive and Corrective Control Applications for Dynamic Security Enhancement in Power Systems
    Genc, Istemihan
    Diao, Ruisheng
    Vittal, Vijay
    Kolluri, Sharma
    Mandal, Sujit
    [J]. IEEE TRANSACTIONS ON POWER SYSTEMS, 2010, 25 (03) : 1611 - 1619
  • [9] Reinforcement Learning for Electric Power System Decision and Control: Past Considerations and Perspectives
    Glavic, Mevludin
    Fonteneau, Raphael
    Ernst, Damien
    [J]. IFAC PAPERSONLINE, 2017, 50 (01): : 6918 - 6927
  • [10] Haarnoja T, 2018, Arxiv, DOI [arXiv:1801.01290, 10.48550/ARXIV.1801.01290]