Dual-Layer Q-Learning Strategy for Energy Management of Battery Storage in Grid-Connected Microgrids

被引:3
作者
Ali, Khawaja Haider [1 ,2 ]
Abusara, Mohammad [1 ]
Tahir, Asif Ali [1 ]
Das, Saptarshi [1 ]
机构
[1] Univ Exeter, Fac Environm Sci & Econ, Penryn Campus, Cornwall TR10 9FE, England
[2] Sukkur IBA Univ, Dept Elect Engn, Sukkur 65200, Pakistan
基金
英国工程与自然科学研究理事会;
关键词
reinforcement learning (RL); microgrid; energy management; offline and online RL; dual-layer Q-learning;
D O I
10.3390/en16031334
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
Real-time energy management of battery storage in grid-connected microgrids can be very challenging due to the intermittent nature of renewable energy sources (RES), load variations, and variable grid tariffs. Two reinforcement learning (RL)-based energy management systems have been previously used, namely, offline and online methods. In offline RL, the agent learns the optimum policy using forecasted generation and load data. Once the convergence is achieved, battery commands are dispatched in real time. The performance of this strategy highly depends on the accuracy of the forecasted data. An agent in online RL learns the best policy by interacting with the system in real time using real data. Online RL deals better with the forecasted error but can take a longer time to converge. This paper proposes a novel dual layer Q-learning strategy to address this challenge. The first (upper) layer is conducted offline to produce directive commands for the battery system for a 24 h horizon. It uses forecasted data for generation and load. The second (lower) Q-learning-based layer refines these battery commands every 15 min by considering the changes happening in the RES and load demand in real time. This decreases the overall operating cost of the microgrid as compared with online RL by reducing the convergence time. The superiority of the proposed strategy (dual-layer RL) has been verified by simulation results after comparing it with individual offline and online RL algorithms.
引用
收藏
页数:17
相关论文
共 25 条
[1]   State-of-the-art in artificial neural network applications: A survey [J].
Abiodun, Oludare Isaac ;
Jantan, Aman ;
Omolara, Abiodun Esther ;
Dada, Kemi Victoria ;
Mohamed, Nachaat AbdElatif ;
Arshad, Humaira .
HELIYON, 2018, 4 (11)
[2]   Reinforcement Learning for Energy-Storage Systems in Grid-Connected Microgrids: An Investigation of Online vs. Offline Implementation [J].
Ali, Khawaja Haider ;
Sigalo, Marvin ;
Das, Saptarshi ;
Anderlini, Enrico ;
Tahir, Asif Ali ;
Abusara, Mohammad .
ENERGIES, 2021, 14 (18)
[3]   Reactive control of a two-body point absorber using reinforcement learning [J].
Anderlini, E. ;
Forehand, D. I. M. ;
Bannon, E. ;
Xiao, Q. ;
Abusara, M. .
OCEAN ENGINEERING, 2018, 148 :650-658
[4]  
Blondin MJ, 2020, P AMER CONTR CONF, P1489, DOI [10.23919/acc45564.2020.9148017, 10.23919/ACC45564.2020.9148017]
[5]   A battery management strategy in microgrid for personalized customer requirements [J].
Chen, Pengzhan ;
Liu, Mengchao ;
Chen, Chuanxi ;
Shang, Xin .
ENERGY, 2019, 189
[6]  
Delgado R., 2012, CRITICAL RACE THEORY, V3
[7]   Residual electricity demand: An empirical investigation [J].
Do, Linh Phuong Catherine ;
Lyocsa, Stefan ;
Molnar, Peter .
APPLIED ENERGY, 2021, 283
[8]  
Eddy Foo Y. S., 2011, 8th International Conference on Power Electronics - ECCE Asia, P2374
[9]   Real-time optimal energy management of microgrid with uncertainties based on deep reinforcement learning [J].
Guo, Chenyu ;
Wang, Xin ;
Zheng, Yihui ;
Zhang, Feng .
ENERGY, 2022, 238
[10]   Impacts of Demand-Side Management on Electrical Power Systems: A Review [J].
Jabir, Hussein Jumma ;
Teh, Jiashen ;
Ishak, Dahaman ;
Abunima, Hamza .
ENERGIES, 2018, 11 (05)