Safe Deep Reinforcement Learning for Power System Operation under Scheduled Unavailability

被引:0
作者
Weiss, Xavier [1 ]
Mohammadi, Saeed [1 ]
Khanna, Parag [1 ]
Hesamzadeh, Mohammad Reza [1 ]
Nordstrom, Lars [1 ]
机构
[1] KTH Royal Inst Technol, Sch Elect Engn & Comp Sci, S-10044 Stockholm, Sweden
来源
2023 IEEE POWER & ENERGY SOCIETY GENERAL MEETING, PESGM | 2023年
基金
瑞典研究理事会;
关键词
Deep reinforcement learning; power system operation; deep learning; safe deep reinforcement learning;
D O I
10.1109/PESGM52003.2023.10252619
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
The electrical grid is a safety -critical system, since incorrect actions taken by a power system operator can result in grid failure and cause harm. For this reason, it is desirable to have an automated power system operator that can reliably take actions that avoid grid failure while fulfilling some objective. Given the existing and growing complexity of power system operation, the choice has often fallen on deep reinforcement learning (DRL) agents for automation, but these are neither explainable nor provably safe. Therefore in this work, the effect of shielding on DRL agent survivability, validation computational time, and convergence are explored. To do this, shielded and unshielded DRL agents are evaluated on a standard IEEE 14 -bus network. Agents are tasked with balancing generation and demand through re dispatch and topology changing actions at a human timescale of 5 minutes. To test survivability under controlled conditions, varying degrees of scheduled unavailability events are introduced which could cause grid failure if unaddressed. Results show improved convergence and generally greater survivability of shielded agents compared with unshielded agents. However, the safety assurances provided by the shield increase computational time. This will require trade-offs or optimizations to make real-time deployment more feasible.
引用
收藏
页数:5
相关论文
共 15 条
  • [1] Alshiekh M, 2018, AAAI CONF ARTIF INTE, P2669
  • [2] Chow Y, 2018, ADV NEUR IN, V31
  • [3] Reinforcement Learning Versus Model Predictive Control: A Comparison on a Power System Problem
    Ernst, Damien
    Glavic, Mevludin
    Capitanescu, Florin
    Wehenkel, Louis
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2009, 39 (02): : 517 - 529
  • [4] Fan Jiameng, 2019, SAFETY GUIDED DEEP R
  • [5] Firesmith Donald G, 2003, TECHNICAL REPORT
  • [6] Harris Andrew T, 2020, AIAA SCITECH 2020 FO, P0386
  • [7] Adaptive Power System Emergency Control Using Deep Reinforcement Learning
    Huang, Qiuhua
    Huang, Renke
    Hao, Weituo
    Tan, Jie
    Fan, Rui
    Huang, Zhenyu
    [J]. IEEE TRANSACTIONS ON SMART GRID, 2020, 11 (02) : 1171 - 1182
  • [8] Marot A, 2021, PROC NEURIPS COMPET, P112
  • [9] Mnih V., 2013, PREPRINT
  • [10] Mohammadi A, 2020, INT CONF ACOUST SPEE, P1001, DOI 10.1109/ICASSP40776.2020.9053685