Real-time energy purchase optimization for a storage-integrated photovoltaic system by deep reinforcement learning

被引:35
作者
Kolodziejczyk, Waldemar [1 ]
Zoltowska, Izabela [1 ]
Cichosz, Pawel [1 ]
机构
[1] Warsaw Univ Technol, Pl Politech 1, PL-00661 Warsaw, Poland
关键词
Smart grid; Energy management; Storage control; Deep reinforcement learning; Q-learning; Neural networks; MANAGEMENT; TECHNOLOGIES; GENERATION;
D O I
10.1016/j.conengprac.2020.104598
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The objective of this article is to minimize the cost of energy purchased on a real-time basis for a storage-integrated photovoltaic (PV) system installed in a microgrid. Under non-linear storage charging/discharging characteristics, as well as uncertain solar energy generation, demands, and market prices, it is a complex task. It requires a proper level of tradeoff between storing too much and too little energy in the battery: future excess PV energy is lost in the former case, and demand is exposed to future high electricity prices in the latter case. We propose a reinforcement learning approach to deal with a non-stationary environment and non-linear storage characteristics. To make this approach applicable, a novel formulation of the decision problem is presented, which focuses on the optimization of grid energy purchases rather than on direct storage control. This limits the complexity of the state and action space, making it possible to achieve satisfactory learning speed and avoid stability issues. Then the Q-learning algorithm combined with a dense deep neural network for function representation is used to learn an optimal decision policy. The algorithm incorporates enhancements that were found to improve learning speed and stability by prior work, such as experience replay, target network, and increasing discount factor. Extensive simulation results performed on real data confirm that our approach is effective and outperforms rule-based heuristics.
引用
收藏
页数:12
相关论文
共 53 条
[1]   Robust optimisation for deciding on real-time flexibility of storage-integrated photovoltaic units controlled by intelligent software agents [J].
Ampatzis, Michail ;
Nguyen, Phuong H. ;
Kamphuis, I. G. ;
van Zwam, Arno .
IET RENEWABLE POWER GENERATION, 2017, 11 (12) :1527-1533
[2]  
Anderson C.W., 1987, P 4 INT WORKSH MACH
[3]  
[Anonymous], 1989, LEARNING DELAYED REW
[4]  
[Anonymous], 1993, P 1993 CONN MOD SUMM
[5]   Review on Energy Storage Systems Control Methods in Microgrids [J].
Arani, A. A. Khodadoost ;
Gharehpetian, G. B. ;
Abedi, M. .
INTERNATIONAL JOURNAL OF ELECTRICAL POWER & ENERGY SYSTEMS, 2019, 107 :745-757
[6]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[7]  
Baird L. C., 1995, ICML 95 P 12 INT C M
[8]  
Barto A. G., 1990, P 1990 CONN MOD SUMM
[9]   NEURONLIKE ADAPTIVE ELEMENTS THAT CAN SOLVE DIFFICULT LEARNING CONTROL-PROBLEMS [J].
BARTO, AG ;
SUTTON, RS ;
ANDERSON, CW .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1983, 13 (05) :834-846
[10]  
Bellman R., 1957, DYNAMIC PROGRAMMING, V1st