Safe multi-agent deep reinforcement learning for decentralized low-carbon operation in active distribution networks and multi-microgrids

被引:0
|
作者
Ye, Tong [1 ,2 ]
Huang, Yuping [1 ,2 ,3 ,4 ]
Yang, Weijia [2 ,3 ,4 ]
Cai, Guotian [1 ,2 ,3 ,4 ]
Yang, Yuyao [5 ]
Pan, Feng [5 ]
机构
[1] Univ Sci & Technol China, Sch Energy Sci & Engn, Hefei 230026, Peoples R China
[2] Chinese Acad Sci, Guangzhou Inst Energy Convers, Guangzhou 510640, Peoples R China
[3] CAS Key Lab Renewable Energy, Guangzhou 510640, Peoples R China
[4] Guangdong Prov Key Lab Renewable Energy, Guangzhou 510640, Peoples R China
[5] Guangdong Power Grid Co Ltd, Metrol Ctr, Qingyuan 511545, Peoples R China
关键词
Active distribution network; Carbon emission allocation; Low-carbon economic operation; Multi-microgrid operation; Safe multi-agent deep reinforcement learning;
D O I
10.1016/j.apenergy.2025.125609
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
Due to fundamental differences in operational entities between distribution networks and microgrids, the equitable allocation of carbon responsibilities remains challenging. Furthermore, achieving real-time, efficient, and secure low-carbon economic dispatch in decentralized multi-entities continues to face obstacles. Therefore, we propose a co-optimization framework for Active Distribution Networks (ADNs) and multi-Microgrids (MMGs) to improve operational efficiency and reduce carbon emissions through adaptive coordination and decisionmaking. To facilitate decentralized low-carbon decision-making, we introduce the Spatiotemporal Carbon Intensity Equalization Method (STCIEM). This method ensures privacy and fairness by processing local data and equitably distributing carbon responsibilities. Additionally, we propose a non-cooperative optimization strategy that enables entities to optimize their operations independently while considering both economic and environmental interests. To address the challenges of real-time decision-making and the non-convex nature of lowcarbon optimization inherent in traditional approaches, we have developed the Enhanced Action Projection Multi-Agent Twin Delayed Deep Deterministic Policy Gradient (EAP-MATD3) algorithm. This algorithm enhances the actor's objective to address the actor-critic mismatch problem, thereby outperforming conventional safe multi-agent deep reinforcement learning methods by generating optimized actions that adhere to physical system constraints. Experiments conducted on the modified IEEE 33-bus network and IEEE 123-bus network demonstrate the superiority of our approach in effectively balancing economic and environmental objectives within complex energy systems.
引用
收藏
页数:21
相关论文
共 50 条
  • [41] Deep Multi-Agent Reinforcement Learning: A Survey
    Liang X.-X.
    Feng Y.-H.
    Ma Y.
    Cheng G.-Q.
    Huang J.-C.
    Wang Q.
    Zhou Y.-Z.
    Liu Z.
    Zidonghua Xuebao/Acta Automatica Sinica, 2020, 46 (12): : 2537 - 2557
  • [42] Multi-agent deep reinforcement learning: a survey
    Sven Gronauer
    Klaus Diepold
    Artificial Intelligence Review, 2022, 55 : 895 - 943
  • [43] Lenient Multi-Agent Deep Reinforcement Learning
    Palmer, Gregory
    Tuyls, Karl
    Bloembergen, Daan
    Savani, Rahul
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS (AAMAS' 18), 2018, : 443 - 451
  • [44] Multi-agent deep reinforcement learning: a survey
    Gronauer, Sven
    Diepold, Klaus
    ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (02) : 895 - 943
  • [45] Safe Multi-Agent Deep Reinforcement Learning for Dynamic Virtual Network Allocation
    Suzuki, Akito
    Harada, Shigeaki
    2020 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2020,
  • [46] A Stochastic Bilevel Model to Manage Active Distribution Networks With Multi-Microgrids
    Toutounchi, Amir Naebi
    Seyedshenava, Seyedjalal
    Contreras, Javier
    Akbarimajd, Adel
    IEEE SYSTEMS JOURNAL, 2019, 13 (04): : 4190 - 4199
  • [47] Deep Reinforcement Learning for Multi-Agent Power Control in Heterogeneous Networks
    Zhang, Lin
    Liang, Ying-Chang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2021, 20 (04) : 2551 - 2564
  • [48] Learning to Communicate with Deep Multi-Agent Reinforcement Learning
    Foerster, Jakob N.
    Assael, Yannis M.
    de Freitas, Nando
    Whiteson, Shimon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [49] Routing with Graph Convolutional Networks and Multi-Agent Deep Reinforcement Learning
    Bhavanasi, Sai Shreyas
    Pappone, Lorenzo
    Esposito, Flavio
    2022 IEEE CONFERENCE ON NETWORK FUNCTION VIRTUALIZATION AND SOFTWARE DEFINED NETWORKS (IEEE NFV-SDN), 2022, : 72 - 77
  • [50] Shield Decentralization for Safe Multi-Agent Reinforcement Learning
    Melcer, Daniel
    Amato, Christopher
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,