Reinforcement learning for robust voltage control in distribution grids under uncertainties

被引:10
作者
Petrusev, Aleksandr [1 ,2 ]
Putratama, Muhammad Andy [1 ]
Rigo-Mariani, Remy [1 ]
Debusschere, Vincent [1 ]
Reignier, Patrick [2 ]
Hadjsaid, Nouredine [1 ]
机构
[1] Univ Grenoble Alpes, CNRS, Grenoble INP, G2Elab, F-38000 Grenoble, France
[2] Univ Grenoble Alpes, CNRS, Grenoble INP, LIG, F-38000 Grenoble, France
关键词
Voltage control; Reinforcement learning; TD3PG; PPO; Flexibility; PV production; Batteries; Distribution grid; Second -order conic relaxation; Optimal power flow;
D O I
10.1016/j.segan.2022.100959
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
Traditional optimization-based voltage controllers for distribution grid applications require consump-tion/production values from the meters as well as accurate grid data (i.e., line impedances) for modeling purposes. Those algorithms are sensitive to uncertainties, notably in consumption and production forecasts or grid models. This paper focuses on the latter. Indeed, line parameters gradually deviate from their original values over time due to exploitation and weather conditions. Also, those data are oftentimes not fully available at the low voltage side thus creating sudden changes between the datasheet and the actual value. To mitigate the impact of uncertain line parameters, this paper proposes the use of a deep reinforcement learning algorithm for voltage regulation purposes in a distribution grid with PV production by controlling the setpoints of distributed storage units as flexibilities. Two algorithms are considered, namely TD3PG and PPO. A two-stage strategy is also proposed, with offline training on a grid model and further online training on an actual system (with distinct impedance values). The controllers' performances are assessed concerning the algo-rithms' hyperparameters, and the obtained results are compared with a second-order conic relaxation optimization-based control. The results show the relevance of the RL-based control, in terms of accuracy, robustness to gradual or sudden variations on the line impedances, and significant speed improvement (once trained). Validation runs are performed on a simple 11-bus system before the method's scalability is tested on a 55-bus network.(c) 2022 Elsevier Ltd. All rights reserved.
引用
收藏
页数:12
相关论文
共 27 条
[1]   Stochastic Distribution System Operation Considering Voltage Regulation Risks in the Presence of PV Generation [J].
Agalgaonkar, Yashodhan P. ;
Pal, Bikash C. ;
Jabr, Rabih A. .
IEEE TRANSACTIONS ON SUSTAINABLE ENERGY, 2015, 6 (04) :1315-1324
[2]  
Cao D, 2021, IEEE Transactions on Smart Grid, P1
[3]   Model-free voltage control of active distribution system with PVs using surrogate model-based deep reinforcement learning [J].
Cao, Di ;
Zhao, Junbo ;
Hu, Weihao ;
Ding, Fei ;
Yu, Nanpeng ;
Huang, Qi ;
Chen, Zhe .
APPLIED ENERGY, 2022, 306
[4]   Real-time prediction of grid voltage and frequency using artificial neural networks: An experimental validation [J].
Chettibi, N. ;
Pavan, A. Massi ;
Mellit, A. ;
Forsyth, A. J. ;
Todd, R. .
SUSTAINABLE ENERGY GRIDS & NETWORKS, 2021, 27
[5]  
data.Jondon, SMART METER ENERGY C
[6]   Deep-Reinforcement-Learning-Based Autonomous Voltage Control for Power Grid Operations [J].
Duan, Jiajun ;
Shi, Di ;
Diao, Ruisheng ;
Li, Haifeng ;
Wang, Zhiwei ;
Zhang, Bei ;
Bian, Desong ;
Yi, Zhehan .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2020, 35 (01) :814-817
[7]  
Fujimoto S, 2018, Arxiv, DOI [arXiv:1802.09477, 10.48550/arXiv.1802.09477]
[8]   Reinforcement learning: A survey [J].
Kaelbling, LP ;
Littman, ML ;
Moore, AW .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 1996, 4 :237-285
[9]  
Keskar N.S., ICLR 2017
[10]  
Kingma DP, 2014, ADV NEUR IN, V27