Online Preventive Control for Transmission Overload Relief Using Safe Reinforcement Learning With Enhanced Spatial-Temporal Awareness

被引:6
|
作者
Cui, Han [1 ]
Ye, Yujian [2 ]
Hu, Jianxiong [2 ]
Tang, Yi [2 ]
Lin, Zizhao [3 ]
Strbac, Goran [4 ]
机构
[1] Southeast Univ, Sch Cyber Sci & Engn, Nanjing 210096, Peoples R China
[2] Southeast Univ, Sch Elect Engn, Nanjing 210096, Peoples R China
[3] Shenzhen Power Supply Bur, Shenzhen 518000, Peoples R China
[4] Imperial Coll London, Dept Elect & Elect Engn, London SW7 2AZ, England
关键词
Optimization; Substations; Generators; Switches; Uncertainty; Indexes; Voltage control; Constrained Markov decision process; online preventive control; power system spatial-temporal perception; safe deep reinforcement learning; transmission overload; GLOBAL SENSITIVITY-ANALYSIS; POLYNOMIAL CHAOS EXPANSION; POWER-SYSTEMS; FLOW;
D O I
10.1109/TPWRS.2023.3257259
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The risk of transmission overload (TO) in power grids is increasing with the large-scale integration of intermittent renewable energy sources. An effective online preventive control schemes proves to be vital in safeguarding the security of power systems. In this paper, we formulate the online preventive control problem for TO alleviation as a constrained Markov decision process (CMDP), targeted to reduce the load rate of overloaded lines by implementing generation re-dispatch, transmission and busbar switching actions. The CMDP is solved with a state-of-the-art safe deep reinforcement learning method, based on the computationally efficient interior-point policy optimization (IPO), which facilitates desirable learning behavior towards constraint satisfaction and policy improvement simultaneously. The performance of IPO method is further improved with an enhanced perception of spatial-temporal correlations in power gird nodal and edge features, combining the strength of edge conditioned convolutional network and long short-term memory network, fostering more effective and robust preventive control policies to be devised. Case studies on a real-world system and a large-scale system validate the superior performance of the proposed method in TO alleviation, constraint handling, uncertainty adaptability and stability preservation, as well as its favorable computational performance, through benchmarking against both model-based and reinforcement learning-based baseline methods.
引用
收藏
页码:517 / 532
页数:16
相关论文
共 7 条
  • [1] Spatial-Temporal Flows-Adaptive Street Layout Control Using Reinforcement Learning
    Ye, Qiming
    Feng, Yuxiang
    Candela, Eduardo
    Escribano Macias, Jose
    Stettler, Marc
    Angeloudis, Panagiotis
    SUSTAINABILITY, 2022, 14 (01)
  • [2] Deep Reinforcement Learning for Voltage Control and Renewable Accommodation Using Spatial-Temporal Graph Information
    Li, Jinhao
    Zhang, Ruichang
    Wang, Hao
    Liu, Zhi
    Lai, Hongyang
    Zhang, Yanru
    IEEE TRANSACTIONS ON SUSTAINABLE ENERGY, 2024, 15 (01) : 249 - 262
  • [3] Dynamic Bus Holding Control Using Spatial-Temporal Data - A Deep Reinforcement Learning Approach
    Zhao, Yuguang
    Chen, Gang
    Ma, Hui
    Zuo, Xingquan
    Ai, Guanqun
    AI 2022: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13728 : 661 - 674
  • [4] Spatial-Temporal Traffic Flow Control on Motorways Using Distributed Multi-Agent Reinforcement Learning
    Kusic, Kresimir
    Ivanjko, Edouard
    Vrbanic, Filip
    Greguric, Martin
    Dusparic, Ivana
    MATHEMATICS, 2021, 9 (23)
  • [5] PATROL: A Velocity Control Framework for Autonomous Vehicle via Spatial-Temporal Reinforcement Learning
    Xu, Zhi
    Liu, Shuncheng
    Wu, Ziniu
    Chen, Xu
    Zeng, Kai
    Zheng, Kai
    Su, Han
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2271 - 2280
  • [6] A Spatial-Temporal Deep Reinforcement Learning Model for Large-Scale Centralized Traffic Signal Control
    Yi, Chenglin
    Wu, Jia
    Ren, Yanyu
    Ran, Yunchuan
    Lou, Yican
    2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, : 275 - 280
  • [7] Three-Stage Inverter-Based Peak Shaving and Volt-VAR Control in Active Distribution Networks Using Online Safe Deep Reinforcement Learning
    Nguyen, Hoang Tien
    Choi, Dae-Hyun
    IEEE TRANSACTIONS ON SMART GRID, 2022, 13 (04) : 3266 - 3277