A Cloud-Edge Collaboration Solution for Distribution Network Reconfiguration Using Multi-Agent Deep Reinforcement Learning

被引:13
作者
Gao, Hongjun [1 ]
Wang, Renjun [1 ]
He, Shuaijia [1 ]
Wang, Lingfeng [2 ]
Liu, Junyong
Chen, Zhe [3 ]
机构
[1] Sichuan Univ, Coll Elect Engn, Chengdu 610065, Peoples R China
[2] Univ Wisconsin, Dept Elect Engn & Comp Sci, Milwaukee, WI 53211 USA
[3] Aalborg Univ, Dept Energy Technol, DK-9220 Aalborg, Denmark
基金
中国国家自然科学基金;
关键词
Batch reinforcement learning; cloud-edge collaboration; distribution network reconfiguration; multi-agent deep reinforcement learning; safe reinforcement learning;
D O I
10.1109/TPWRS.2023.3296463
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Network reconfiguration can maintain the optimal operation of distribution network with increasing penetration of distributed generations (DGs). However, network reconfiguration problems may not be solved quickly by traditional methods in large-scale distribution networks. In this context, a cloud-edge collaboration framework based on multi-agent deep reinforcement learning (MADRL) is proposed, where the MADRL model can be trained centrally in the cloud center and decentrally executed in edge servers to reduce the training cost and execution latency of MADRL. In addition, a discrete multi-agent soft actor-critic algorithm (MASAC) is introduced as the basic algorithm to address the non-stationary environment problem in MADRL. Then, online safe learning and offline safe learning are combined for the distribution network reconfiguration task in practice to update the neural networks of MADRL under constraints. Specifically, a novel offline algorithm called multi-agent constraints penalized Q-learning (MACPQ) is proposed to reduce the cost of trial-and-error process of MADRL while allowing agents to pre-train their policies from a historical dataset considering constraints. Meanwhile, a new online MADRL method called primal-dual MASAC is proposed to further improve the performance of agents by directly interacting with the physical distribution network under the safe action exploration. Finally, the superiority of the proposed methods is verified in IEEE 33-bus system and a practical 445-bus system.
引用
收藏
页码:3867 / 3879
页数:13
相关论文
共 37 条
[1]  
Achiam J, 2017, PR MACH LEARN RES, V70
[2]   Real-time operation of distribution network: A deep reinforcement learning-based reconfiguration approach [J].
Bui, Van-Hai ;
Su, Wencong .
SUSTAINABLE ENERGY TECHNOLOGIES AND ASSESSMENTS, 2022, 50
[3]   A Survey on Edge and Edge-Cloud Computing Assisted Cyber-Physical Systems [J].
Cao, Kun ;
Hu, Shiyan ;
Shi, Yang ;
Colombo, Armando ;
Karnouskos, Stamatis ;
Li, Xin .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (11) :7806-7819
[4]   Comprehensive Cost Minimization in Distribution Networks Using Segmented-Time Feeder Reconfiguration and Reactive Power Control of Distributed Generators [J].
Chen, Shuheng ;
Hu, Weihao ;
Chen, Zhe .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2016, 31 (02) :983-993
[5]  
Christodoulou P, 2019, Arxiv, DOI arXiv:1910.07207
[6]   A Win-Win Mode: The Complementary and Coexistence of 5G Networks and Edge Computing [J].
Fang, Fang ;
Wu, Xiaolun .
IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (06) :3983-4003
[7]  
Fujimoto S, 2019, PR MACH LEARN RES, V97
[8]  
Fujimoto Scott, 2019, arXiv, DOI DOI 10.48550/ARXIV.1910.01708
[9]   Batch-Constrained Reinforcement Learning for Dynamic Distribution Network Reconfiguration [J].
Gao, Yuanqi ;
Wang, Wei ;
Shi, Jie ;
Yu, Nanpeng .
IEEE TRANSACTIONS ON SMART GRID, 2020, 11 (06) :5357-5369
[10]   A new heuristic reconfiguration algorithm for large distribution systems [J].
Gomes, FV ;
Carneiro, S ;
Pereira, JLR ;
Vinagre, MP ;
Garcia, PAN ;
Araujo, LR .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2005, 20 (03) :1373-1378