An autonomous control technology based on deep reinforcement learning for optimal active power dispatch

被引:26
作者
Han, Xiaoyun [1 ]
Mu, Chaoxu [1 ]
Yan, Jun [2 ]
Niu, Zeyuan [3 ]
机构
[1] Tianjin Univ, Sch Elect & Informat Engn, Tianjin, Peoples R China
[2] Concordia Univ, Concordia Inst Informat Syst Engn CIISE, Montreal, PQ, Canada
[3] Southeast Univ, Sch Informat Sci & Engn, Nanjing, Peoples R China
关键词
Active power dispatch; Renewable energy penetration; Soft actor-critic (SAC); Imitation learning (IL); Lagrange multiplier method; Robustness;
D O I
10.1016/j.ijepes.2022.108686
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The large-scale renewable energy integration has brought challenges to energy management in modern power systems. Due to the strong randomness and volatility of renewable energy, traditional model-based methods may become insufficient for optimal active power dispatch. To tackle the challenge, this paper proposes an autonomous control method based on soft actor-critic (SAC), a deep-reinforcement learning (DRL) strategy recently developed, which provides an optimal solution for active power dispatch without a mathematical model while improving the renewable energy consumption rate under stable operation. A Lagrange multiplier is introduced to the SAC (LM-SAC) to promote algorithm performance in optimal active power dispatch. A pre-trained scheme based on imitation learning (IL-SAC) is also designed to further improve the training efficiency and robustness of the DRL agent. Simulations on the IEEE 118-bus system with the open platform Grid2Op verify that the proposed algorithm effectively achieves better renewable energy consumption rate and robustness compared with existing DRL algorithms.
引用
收藏
页数:10
相关论文
共 31 条
[1]   Q-Learning-Based Damping Control of Wide-Area Power Systems Under Cyber Uncertainties [J].
Duan, Jiajun ;
Xu, Hao ;
Liu, Wenxin .
IEEE TRANSACTIONS ON SMART GRID, 2018, 9 (06) :6408-6418
[2]   A composite framework coupling multiple feature selection, compound prediction models and novel hybrid swarm optimizer-based synchronization optimization strategy for multi-step ahead short-term wind speed forecasting [J].
Fu, Wenlong ;
Wang, Kai ;
Tan, Jiawen ;
Zhang, Kai .
ENERGY CONVERSION AND MANAGEMENT, 2020, 205
[3]   Deep learning for time series forecasting: The electric load case [J].
Gasparin, Alberto ;
Lukovic, Slobodan ;
Alippi, Cesare .
CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2022, 7 (01) :1-25
[4]  
Haarnoja T, 2018, COMPUT RES REPOS
[5]  
Haarnoja T, 2018, PR MACH LEARN RES, V80
[6]   A Deep Reinforcement Learning-Based Multi-Agent Framework to Enhance Power System Resilience Using Shunt Resources [J].
Kamruzzaman, Md. ;
Duan, Jiajun ;
Shi, Di ;
Benidris, Mohammed .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2021, 36 (06) :5525-5536
[7]   Dynamic Energy Dispatch Based on Deep Reinforcement Learning in IoT-Driven Smart Isolated Microgrids [J].
Lei, Lei ;
Tan, Yue ;
Dahlenburg, Glenn ;
Xiang, Wei ;
Zheng, Kan .
IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (10) :7938-7953
[8]  
Marot A, 2021, NeurIPS 2020 Competition and Demonstration Track, P112
[9]  
Marot A, 2020, NEURIPS2020 CHALLENG
[10]   A Reliability Perspective of the Smart Grid [J].
Moslehi, Khosrow ;
Kumar, Ranjit .
IEEE TRANSACTIONS ON SMART GRID, 2010, 1 (01) :57-64