Multi-Agent Safe Policy Learning for Power Management of Networked Microgrids

被引:101
作者
Zhang, Qianzhi [1 ]
Dehghanpour, Kaveh [1 ]
Wang, Zhaoyu [1 ]
Qiu, Feng [2 ]
Zhao, Dongbo [2 ]
机构
[1] Iowa State Univ, Dept Elect & Comp Engn, Ames, IA 50011 USA
[2] Argonne Natl Lab, Div Energy Syst, Lemont, IL 60439 USA
基金
美国国家科学基金会;
关键词
Training; Power system management; Optimization; Reactive power; Computational modeling; Safety; Indexes; Safe policy learning; multi-agent framework; networked microgrids; power management; policy gradient; ENERGY MANAGEMENT; SYSTEM; STORAGE;
D O I
10.1109/TSG.2020.3034827
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
This article presents a supervised multi-agent safe policy learning (SMAS-PL) method for optimal power management of networked microgrids (MGs) in distribution systems. While unconstrained reinforcement learning (RL) algorithms are black-box decision models that could fail to satisfy grid operational constraints, our proposed method considers AC power flow equations and other operational limits. Accordingly, the training process employs the gradient information of operational constraints to ensure that the optimal control policy functions generate safe and feasible decisions. Furthermore, we have developed a distributed consensus-based optimization approach to train the agents' policy functions while maintaining MGs' privacy and data ownership boundaries. After training, the learned optimal policy functions can be safely used by the MGs to dispatch their local resources, without the need to solve a complex optimization problem from scratch. Numerical experiments have been devised to verify the performance of the proposed method.
引用
收藏
页码:1048 / 1062
页数:15
相关论文
共 35 条
[11]   Distributed Constrained Optimization by Consensus-Based Primal-Dual Perturbation Method [J].
Chang, Tsung-Hui ;
Nedic, Angelia ;
Scaglione, Anna .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2014, 59 (06) :1524-1538
[12]   Reinforcement-Learning-Based Optimal Control of Hybrid Energy Storage Systems in Hybrid AC-DC Microgrids [J].
Duan, Jiajun ;
Yi, Zhehan ;
Shi, Di ;
Lin, Chang ;
Lu, Xiao ;
Wang, Zhiwei .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (09) :5355-5364
[13]   Sensitivity-Based Discrete Coordinate-Descent for Volt/VAr Control in Distribution Networks [J].
Jabr, Rabih A. ;
Dzafic, Izudin .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2016, 31 (06) :4670-4678
[14]  
Kersting W, 2001, 2001 IEEE POWER ENGINEERING SOCIETY WINTER MEETING, CONFERENCE PROCEEDINGS, VOLS 1-3, P908, DOI 10.1109/PESW.2001.916993
[15]   Dynamic Pricing and Energy Consumption Scheduling With Reinforcement Learning [J].
Kim, Byung-Gook ;
Zhang, Yu ;
van der Schaar, Mihaela ;
Lee, Jang-Won .
IEEE TRANSACTIONS ON SMART GRID, 2016, 7 (05) :2187-2198
[16]   Constrained EV Charging Scheduling Based on Safe Deep Reinforcement Learning [J].
Li, Hepeng ;
Wan, Zhiqiang ;
He, Haibo .
IEEE TRANSACTIONS ON SMART GRID, 2020, 11 (03) :2427-2439
[17]  
Li J., 2012, JOURNALSPRINGER, P553, DOI [DOI 10.1007/978-3-642-30223-7_87, 10.1007/978-3-642-30223-7_87]
[18]   Networked Microgrids for Enhancing the Power System Resilience [J].
Li, Zhiyi ;
Shahidehpour, Mohammad ;
Aminifar, Farrokh ;
Alabdulwahab, Ahmed ;
Al-Turki, Yusuf .
PROCEEDINGS OF THE IEEE, 2017, 105 (07) :1289-1310
[19]   Reinforcement Learning-Based Microgrid Energy Trading With a Reduced Power Plant Schedule [J].
Lu, Xiaozhen ;
Xiao, Xingyu ;
Xiao, Liang ;
Dai, Canhuang ;
Peng, Mugen ;
Poor, H. Vincent .
IEEE INTERNET OF THINGS JOURNAL, 2019, 6 (06) :10728-10737
[20]   Active distribution networks planning with integration of demand response [J].
Mokryani, Geev .
SOLAR ENERGY, 2015, 122 :1362-1370