Safe Deep Reinforcement Learning-Based Real-Time Operation Strategy in Unbalanced Distribution System

被引:0
|
作者
Yoon, Yeunggurl [1 ]
Yoon, Myungseok [1 ]
Zhang, Xuehan [2 ]
Choi, Sungyun [1 ]
机构
[1] Korea Univ, Sch Elect Engn, Seoul 02841, South Korea
[2] Fuzhou Univ, Coll Elect Engn & Automat, Fuzhou 350116, Peoples R China
基金
新加坡国家研究基金会;
关键词
Optimization; Mathematical models; Voltage control; Real-time systems; Systems operation; Reactive power; Uncertainty; Deep reinforcement learning; hybrid optimization; quadratic programming; safe deep reinforcement learning; voltage unbalance factor;
D O I
10.1109/TIA.2024.3446735
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Unbalanced voltages are one of the voltage quality issues affecting customer devices in distribution systems. Conventional optimization methods are time-consuming to mitigate unbalanced voltage in real time because these approaches must solve each scenario after observation. Deep reinforcement learning (DRL) is effectively trained offline for real-time operations that overcome the time-consumption problem in practical implementation. This paper proposes a safe deep reinforcement learning (SDRL) based distribution system operation method to mitigate unbalanced voltage for real-time operation and satisfy operational constraints. The proposed SDRL method incorporates a learning module (LM) and a constraint module (CM), controlling the energy storage system (ESS) to improve voltage balancing. The proposed SDRL method is compared with the hybrid optimization (HO) and typical DRL models regarding time consumption and voltage unbalance mitigation. For this purpose, the models operate in modified IEEE-13 node and IEEE-123 node test feeders.
引用
收藏
页码:8273 / 8283
页数:11
相关论文
共 50 条
  • [1] Real-time operation of distribution network: A deep reinforcement learning-based reconfiguration approach
    Bui, Van-Hai
    Su, Wencong
    SUSTAINABLE ENERGY TECHNOLOGIES AND ASSESSMENTS, 2022, 50
  • [2] Deep Reinforcement Learning-Based Dynamic Droop Control Strategy for Real-Time Optimal Operation and Frequency Regulation
    Lee, Woon-Gyu
    Kim, Hak-Man
    IEEE TRANSACTIONS ON SUSTAINABLE ENERGY, 2025, 16 (01) : 284 - 294
  • [3] A Deep Learning-Based Real-time Seizure Detection System
    Shawki, N.
    Elseify, T.
    Cap, T.
    Shah, V
    Obeid, I
    Picone, J.
    2020 IEEE SIGNAL PROCESSING IN MEDICINE AND BIOLOGY SYMPOSIUM, 2020,
  • [4] Real-time Dispatch Strategy for Electric Vehicles Based on Deep Reinforcement Learning
    Li H.
    Li G.
    Wang K.
    Dianli Xitong Zidonghua/Automation of Electric Power Systems, 2020, 44 (22): : 161 - 167
  • [5] Deep Reinforcement Learning-based Real-time Online Energy Management Strategy for Electric Vehicle Charging Stations
    Yang, Zhaoqiang
    Li, Longtan
    Yao, Rui
    Liu, Chunxiu
    Liu, Yimin
    Zhou, Zaiyan
    2024 4TH POWER SYSTEM AND GREEN ENERGY CONFERENCE, PSGEC 2024, 2024, : 490 - 494
  • [6] Towards safe and sustainable reinforcement learning for real-time strategy games
    Andersen, Per-Arne
    Goodwin, Morten
    Granmo, Ole-Christoffer
    INFORMATION SCIENCES, 2024, 679
  • [7] Safe Deep Reinforcement Learning-Based Real-Time Multi-Energy Management in Combined Heat and Power Microgrids
    Hu, Bo
    Gong, Yuzhong
    Liang, Xiaodong
    Chung, C. Y.
    Noble, Bram F.
    Poelzer, Greg
    IEEE ACCESS, 2024, 12 : 193581 - 193593
  • [8] Real-time Operation Optimization in Active Distribution Networks Based on Multi-agent Deep Reinforcement Learning
    Jie Xu
    Hongjun Gao
    Renjun Wang
    Junyong Liu
    Journal of Modern Power Systems and Clean Energy, 2024, 12 (03) : 886 - 899
  • [9] Real-Time Operation Optimization in Active Distribution Networks Based on Multi-Agent Deep Reinforcement Learning
    Xu, Jie
    Gao, Hongjun
    Wang, Renjun
    Liu, Junyong
    JOURNAL OF MODERN POWER SYSTEMS AND CLEAN ENERGY, 2024, 12 (03) : 886 - 899
  • [10] A deep reinforcement learning-based scheduling framework for real-time workflows in the cloud environment
    Pan, Jiahui
    Wei, Yi
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 255