Learning to Operate Distribution Networks With Safe Deep Reinforcement Learning

被引:79
作者
Li, Hepeng [1 ]
He, Haibo [1 ]
机构
[1] Univ Rhode Isl, Dept Elect Comp & Biomed Engn, Kingston, RI 02881 USA
基金
美国国家科学基金会;
关键词
Uncertainty; Distribution networks; Voltage control; Reactive power; Optimization; Reinforcement learning; Load modeling; Distribution systems; safe deep reinforcement learning; constrained Markov decision process (MDP); mixed discrete and continuous actions; data-driven decision making; MANAGEMENT;
D O I
10.1109/TSG.2022.3142961
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
In this paper, we propose a safe deep reinforcement learning (SDRL) based method to solve the problem of optimal operation of distribution networks (OODN). We formulate OODN as a constrained Markov decision process (CMDP). The objective is to achieve adaptive voltage regulation and energy cost minimization considering the uncertainty of renewable resources (RSs), nodal loads and energy prices. The control actions include the number of in-operation units of the switchable capacitor banks (SCBs), the tap position of the on-load tap-changers (OLTCs) and voltage regulators (VRs), the active and reactive power of distributed generators (DGs), and the charging and discharging power of battery storage systems (BSSs). To optimize the discrete and continuous actions simultaneously, a stochastic policy built upon a joint distribution of mixed random variables is designed and learned through a neural network approximator. To guarantee that safety constraints are satisfied, constrained policy optimization (CPO) is employed to train the neural network. The proposed approach enables the agent to learn a cost-effective operating strategy through exploring safe scheduling actions. Compared to traditional deep reinforcement learning (DRL) methods that allow agents to freely explore any behaviors during training, the proposed approach is more practical to be applied in a real system. Simulation results on a modified IEEE-34 node system and a modified IEEE-123 node system demonstrate the effectiveness of the proposed method.
引用
收藏
页码:1860 / 1872
页数:13
相关论文
共 33 条
[1]  
Achiam J, 2017, PR MACH LEARN RES, V70
[2]  
[Anonymous], 2004, Electric Power Distribution handbook
[3]  
California ISO Open Access Same-Time Inf. Syst, 2021, OASIS
[4]   A Multi-Agent Deep Reinforcement Learning Based Voltage Regulation Using Coordinated PV Inverters [J].
Cao, Di ;
Hu, Weihao ;
Zhao, Junbo ;
Huang, Qi ;
Chen, Zhe ;
Blaabjerg, Frede .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2020, 35 (05) :4120-4123
[5]   A Comprehensive Centralized Approach for Voltage Constraints Management in Active Distribution Grid [J].
Capitanescu, Florin ;
Bilibin, Ilya ;
Romero Ramos, Esther .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2014, 29 (02) :933-942
[6]   A Two-Stage Robust Reactive Power Optimization Considering Uncertain Wind Power Integration in Active Distribution Networks [J].
Ding, Tao ;
Liu, Shiyu ;
Yuan, Wei ;
Bie, Zhaohong ;
Zeng, Bo .
IEEE TRANSACTIONS ON SUSTAINABLE ENERGY, 2016, 7 (01) :301-311
[7]   Reinforcement Learning Approach for Optimal Distributed Energy Management in a Microgrid [J].
Foruzan, Elham ;
Soh, Leen-Kiat ;
Asgarpoor, Sohrab .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2018, 33 (05) :5749-5758
[8]   Integrated Day-Ahead Scheduling Considering Active Management in Future Smart Distribution System [J].
Gao, Hongjun ;
Wang, Lingfeng ;
Liu, Junyong ;
Wei, Zhenbo .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2018, 33 (06) :6049-6061
[9]  
Haarnoja T., 2018, PR MACH LEARN RES
[10]   Distributionally Robust Optimal Power Flow in Multi-Microgrids With Decomposition and Guaranteed Convergence [J].
Huang, Wanjun ;
Zheng, Weiye ;
Hill, David J. .
IEEE TRANSACTIONS ON SMART GRID, 2021, 12 (01) :43-55