Dynamic Energy Dispatch Based on Deep Reinforcement Learning in IoT-Driven Smart Isolated Microgrids

被引:73
作者
Lei, Lei [1 ]
Tan, Yue [2 ]
Dahlenburg, Glenn [3 ]
Xiang, Wei [4 ]
Zheng, Kan [2 ]
机构
[1] Univ Guelph, Coll Engn & Phys Sci, Guelph, ON N1G 2W1, Canada
[2] Beijing Univ Posts & Telecommun, Intelligent Comp & Commun Lab, Key Lab Universal Wireless Commun, Minist Educ, Beijing 100876, Peoples R China
[3] Ergon Energy, Future Networks, Cairns, Qld 4870, Australia
[4] La Trobe Univ, Sch Engn & Math Sci, Melbourne, Vic 3086, Australia
关键词
Stochastic processes; Energy management; Uncertainty; Batteries; Predictive models; Internet of Things; Optimal control; Deep reinforcement learning (DRL); energy management; Internet of Things (IoT); microgrid; MODEL-PREDICTIVE CONTROL; STOCHASTIC OPTIMIZATION; MANAGEMENT-SYSTEM; OPERATION; TUTORIAL; STRATEGY; STORAGE;
D O I
10.1109/JIOT.2020.3042007
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Microgrids (MGs) are small, local power grids that can operate independently from the larger utility grid. Combined with the Internet of Things (IoT), a smart MG can leverage the sensory data and machine learning techniques for intelligent energy management. This article focuses on deep reinforcement learning (DRL)-based energy dispatch for IoT-driven smart isolated MGs with diesel generators (DGs), photovoltaic (PV) panels, and a battery. A finite-horizon partial observable Markov decision process (POMDP) model is formulated and solved by learning from historical data to capture the uncertainty in future electricity consumption and renewable power generation. In order to deal with the instability problem of DRL algorithms and unique characteristics of finite-horizon models, two novel DRL algorithms, namely, finite-horizon deep deterministic policy gradient (FH-DDPG) and finite-horizon recurrent deterministic policy gradient (FH-RDPG), are proposed to derive energy dispatch policies with and without fully observable state information. A case study using real isolated MG data is performed, where the performance of the proposed algorithms are compared with the other baseline DRL and non-DRL algorithms. Moreover, the impact of uncertainties on MG performance is decoupled into two levels and evaluated, respectively.
引用
收藏
页码:7938 / 7953
页数:16
相关论文
共 46 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]   Review of Internet of Things (IoT) in Electric Power and Energy Systems [J].
Bedi, Guneet ;
Venayagamoorthy, Ganesh Kumar ;
Singh, Rajendra ;
Brooks, Richard R. ;
Wang, Kuang-Ching .
IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (02) :847-870
[3]  
Bergveld HendrikJohannes., 2001, Battery Management Systems - Design by Modelling
[4]  
Duan Y, 2016, PR MACH LEARN RES, V48
[5]   A Stochastic Multi-Objective Framework for Optimal Scheduling of Energy Storage Systems in Microgrids [J].
Farzin, Hossein ;
Fotuhi-Firuzabad, Mahmud ;
Moeini-Aghtaie, Moein .
IEEE TRANSACTIONS ON SMART GRID, 2017, 8 (01) :117-127
[6]   Reinforcement Learning Approach for Optimal Distributed Energy Management in a Microgrid [J].
Foruzan, Elham ;
Soh, Leen-Kiat ;
Asgarpoor, Sohrab .
IEEE TRANSACTIONS ON POWER SYSTEMS, 2018, 33 (05) :5749-5758
[7]  
Francois-Lavet V., 2016, PROC EUR WORKSHOP RE
[8]   Microgrids Energy Management Using Robust Convex Programming [J].
Giraldo, Juan S. ;
Castrillon, Jhon A. ;
Lopez, Juan Camilo ;
Rider, Marcos J. ;
Castro, Carlos A. .
IEEE TRANSACTIONS ON SMART GRID, 2019, 10 (04) :4520-4530
[9]  
Grondman I., 2013, 2013 INT JOINT C NEU, P1
[10]  
Heess N, 2015, ARXIV151204455CS