Real-time energy purchase optimization for a storage-integrated photovoltaic system by deep reinforcement learning

被引:33
作者
Kolodziejczyk, Waldemar [1 ]
Zoltowska, Izabela [1 ]
Cichosz, Pawel [1 ]
机构
[1] Warsaw Univ Technol, Pl Politech 1, PL-00661 Warsaw, Poland
关键词
Smart grid; Energy management; Storage control; Deep reinforcement learning; Q-learning; Neural networks; MANAGEMENT; TECHNOLOGIES; GENERATION;
D O I
10.1016/j.conengprac.2020.104598
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The objective of this article is to minimize the cost of energy purchased on a real-time basis for a storage-integrated photovoltaic (PV) system installed in a microgrid. Under non-linear storage charging/discharging characteristics, as well as uncertain solar energy generation, demands, and market prices, it is a complex task. It requires a proper level of tradeoff between storing too much and too little energy in the battery: future excess PV energy is lost in the former case, and demand is exposed to future high electricity prices in the latter case. We propose a reinforcement learning approach to deal with a non-stationary environment and non-linear storage characteristics. To make this approach applicable, a novel formulation of the decision problem is presented, which focuses on the optimization of grid energy purchases rather than on direct storage control. This limits the complexity of the state and action space, making it possible to achieve satisfactory learning speed and avoid stability issues. Then the Q-learning algorithm combined with a dense deep neural network for function representation is used to learn an optimal decision policy. The algorithm incorporates enhancements that were found to improve learning speed and stability by prior work, such as experience replay, target network, and increasing discount factor. Extensive simulation results performed on real data confirm that our approach is effective and outperforms rule-based heuristics.
引用
收藏
页数:12
相关论文
共 53 条
  • [1] Robust optimisation for deciding on real-time flexibility of storage-integrated photovoltaic units controlled by intelligent software agents
    Ampatzis, Michail
    Nguyen, Phuong H.
    Kamphuis, I. G.
    van Zwam, Arno
    [J]. IET RENEWABLE POWER GENERATION, 2017, 11 (12) : 1527 - 1533
  • [2] Anderson C. W., 1987, P 4 INT WORKSH MACH
  • [3] [Anonymous], 1989, (Ph.D. thesis
  • [4] [Anonymous], 1993, P 1993 CONN MOD SUMM
  • [5] [Anonymous], 2012, ar**v preprint ar**v:1207.0580
  • [6] Review on Energy Storage Systems Control Methods in Microgrids
    Arani, A. A. Khodadoost
    Gharehpetian, G. B.
    Abedi, M.
    [J]. INTERNATIONAL JOURNAL OF ELECTRICAL POWER & ENERGY SYSTEMS, 2019, 107 : 745 - 757
  • [7] Deep Reinforcement Learning A brief survey
    Arulkumaran, Kai
    Deisenroth, Marc Peter
    Brundage, Miles
    Bharath, Anil Anthony
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) : 26 - 38
  • [8] Baird L. C., 1995, ICML 95 P 12 INT C M
  • [9] Barto A. G., 1990, P 1990 CONN MOD SUMM
  • [10] NEURONLIKE ADAPTIVE ELEMENTS THAT CAN SOLVE DIFFICULT LEARNING CONTROL-PROBLEMS
    BARTO, AG
    SUTTON, RS
    ANDERSON, CW
    [J]. IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1983, 13 (05): : 834 - 846