Decoupled Volt/Var Control With Safe Reinforcement Learning Based on Approximate Bayesian Inference

被引:1
作者
Zhang, Yang [1 ]
Wang, Peng [1 ]
Yu, Liying [1 ]
Li, Ning [1 ]
Qiu, Jing [2 ]
机构
[1] Shanghai Jiao Tong Univ, Dept Automation, Shanghai 200240, Peoples R China
[2] Univ Sydney, Sch Elect & Informat Engn, Sydney, NSW 2006, Australia
基金
中国国家自然科学基金;
关键词
Voltage control; Costs; Reactive power; Optimization; Safety; Trajectory; Training; Bayes methods; Uncertainty; Power system stability; Active distribution networks; decoupled volt/var control; approximate Bayesian inference; twin-critic safe reinforcement learning; reactive power fine-tuning strategy; DISTRIBUTION NETWORKS;
D O I
10.1109/TSTE.2024.3485060
中图分类号
X [环境科学、安全科学];
学科分类号
08 ; 0830 ;
摘要
Deep reinforcement learning (DRL) has proven promising for addressing the problems brought by the uncertainties of renewable energy sources in the volt/var control (VVC) problem of active distribution networks (ADNs). However, deploying a conventional DRL-based VVC model cannot simultaneously achieve optimal reactive power generation loss (RPGL) and satisfy voltage safety constraints (VSCs). To tackle the challenges, we propose an approximate Bayesian inference (ABI)-based twin-critic safe reinforcement learning (TCSRL) algorithm with a reactive power fine-tuning (RPFT) strategy. The cost optimization of violating VSCs and RPGL optimization is first decoupled to obtain a better-performing VVC policy by formulating VVC as an ABI problem. A novel TCSRL algorithm is then developed to address the ABI problem. In the algorithm, the trajectory distribution satisfying the cost threshold constraint is analytically computed with an optimality guarantee, and the VVC policy is learned toward lower RPGL. In addition, twin-critic networks are constructed to improve the algorithm robustness in estimating RPGL and the cost of violating VSCs. An RPFT strategy is finally modeled and integrated into the TCSRL algorithm to ensure a 100% voltage safety rate in the training and verification phases. The performance of the proposed algorithms is validated on different IEEE bus systems.
引用
收藏
页码:797 / 811
页数:15
相关论文
共 31 条
[1]  
Abdolmaleki A., 2018, ARXIV
[2]   A Framework for Volt-VAR Optimization in Distribution Systems [J].
Ahmadi, Hamed ;
Marti, Jose R. ;
Dommel, Hermann W. .
IEEE TRANSACTIONS ON SMART GRID, 2015, 6 (03) :1473-1483
[3]   Robust Volt-Var control of a smart distribution system under uncertain voltage-dependent load and renewable production [J].
Azarnia, Mahsa ;
Rahimiyan, Morteza .
INTERNATIONAL JOURNAL OF ELECTRICAL POWER & ENERGY SYSTEMS, 2022, 134
[4]  
BOX GEP, 1973, BAYESIAN INFERENCE S
[5]   Distributional Soft Actor-Critic: Off-Policy Reinforcement Learning for Addressing Value Estimation Errors [J].
Duan, Jingliang ;
Guan, Yang ;
Li, Shengbo Eben ;
Ren, Yangang ;
Sun, Qi ;
Cheng, Bo .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (11) :6584-6598
[6]   Risk-Aware Operating Regions for PV-Rich Distribution Networks Considering Irradiance Variability [J].
Duque, Edgar Mauricio Salazar ;
Giraldo, Juan S. ;
Vergara, Pedro P. ;
Nguyen, Phuong H. ;
van der Molen, Anne ;
Slootweg, J. G. .
IEEE TRANSACTIONS ON SUSTAINABLE ENERGY, 2023, 14 (04) :2092-2108
[7]   MPC-Based Coordinated Voltage Control in Active Distribution Networks Incorporating CVR and DR [J].
Dutta, Arunima ;
Ganguly, Sanjib ;
Kumar, Chandan .
IEEE TRANSACTIONS ON INDUSTRY APPLICATIONS, 2022, 58 (04) :4309-4318
[8]   A Physics-Constrained TD3 Algorithm for Simultaneous Virtual Inertia and Damping Control of Grid-Connected Variable Speed DFIG Wind Turbines [J].
Egbomwan, Osarodion Emmanuel ;
Chaoui, Hicham ;
Liu, Shichao .
IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 :958-969
[9]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[10]   Model-augmented safe reinforcement learning for Volt-VAR control in power distribution networks [J].
Gao, Yuanqi ;
Yu, Nanpeng .
APPLIED ENERGY, 2022, 313