Deep Reinforcement Learning-Based Tie-Line Power Adjustment Method for Power System Operation State Calculation

被引:11
作者
Xu, Huating [1 ]
Yu, Zhihong [1 ]
Zheng, Qingping [2 ]
Hou, Jinxiu [1 ]
Wei, Yawei [1 ]
Zhang, Zhijian [3 ]
机构
[1] China Elect Power Res Inst, Beijing 100192, Peoples R China
[2] Beijing Univ Posts & Telecommun, Sch Informat & Commun Engn, Beijing 100876, Peoples R China
[3] State Grid Beijing Elect Power Dispatching & Cont, Beijing 100031, Peoples R China
关键词
Load flow; Generators; Power system stability; Reinforcement learning; Power grids; Training; Operation state calculation; tie-line power adjustment; deep reinforcement learning; stepwise training; prioritized target replay; NEURAL-NETWORKS; WIND POWER;
D O I
10.1109/ACCESS.2019.2949480
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Operation state calculation (OSC) provides safe operating boundaries for power systems. The operators rely on the software-aid OSC results to dispatch the generators for grid control. Currently, the OSC workload has increased dramatically, as the power grid structure expands rapidly to mitigate renewable source integration. However, the OSC is processed with a lot of manual interventions in most dispatching centers, which makes the OSC error-prone and personnel-experience oriented. Therefore, it is crucial to upgrade the current OSC in an automatic mode for efficiency and quality improvements. An essential process in the OSC is the tie-line power (TP) adjustment. In this paper, a new TP adjustment method is proposed using an adaptive mapping strategy and a Markov Decision Process (MDP) formulation. Then, a model-free deep reinforcement learning (DRL) algorithm is proposed to solve the formulated MDP and learn an optimal adjustment strategy. The improvement techniques of stepwise training and prioritized target replay are included to decompose the large-scale complex problems and improve the training efficiency. Finally, five experiments are conducted on the IEEE 39-bus system and an actual 2725-bus power grid of China for the effectiveness demonstration.
引用
收藏
页码:156160 / 156174
页数:15
相关论文
共 36 条
[1]   Convolutional Neural Networks for Automatic State-Time Feature Extraction in Reinforcement Learning Applied to Residential Load Control [J].
Claessens, Bert J. ;
Vrancx, Peter ;
Ruelens, Frederik .
IEEE TRANSACTIONS ON SMART GRID, 2018, 9 (04) :3259-3269
[2]  
Dewey D., 2014, 2014 AAAI Spring Symposium Series
[3]   The Role of Concentrating Solar Power Toward High Renewable Energy Penetrated Power Systems [J].
Du, Ershun ;
Zhang, Ning ;
Hodge, Bri-Mathias ;
Wang, Qin ;
Kang, Chongqing ;
Kroposki, Benjamin ;
Xia, Qing .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2018, 33 (06) :6630-6641
[4]   Introducing Uncertainty Components in Locational Marginal Prices for Pricing Wind Power and Load Uncertainties [J].
Fang, Xin ;
Hodge, Bri-Mathias ;
Du, Ershun ;
Kang, Chongqing ;
Li, Fangxing .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2019, 34 (03) :2013-2024
[5]   Dynamic Voltage and Frequency Scaling in NoCs with Supervised and Reinforcement Learning Techniques [J].
Fettes, Quintin ;
Clark, Mark ;
Bunescu, Razvan ;
Karanth, Avinash ;
Louri, Ahmed .
IEEE TRANSACTIONS ON COMPUTERS, 2019, 68 (03) :375-389
[6]  
Hadfield-Menell D, 2017, ADV NEUR IN, V30
[7]  
Kingma DP, 2014, ARXIV
[8]  
Krizhevsky A., 2010, Unpublished Manuscript, V40, P1
[9]   Generation Adjustment Method Based on Bus-Dependent Participation Factor [J].
Lee, Jin-Oh ;
Kim, Yun-Su ;
Kim, Eung-Sang ;
Moon, Seung-Il .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2018, 33 (02) :1959-1969
[10]  
Lillicrap TP, 2015, ARXIV150902971