On-Line Building Energy Optimization Using Deep Reinforcement Learning

被引:393
|
作者
Mocanu, Elena [1 ,2 ]
Mocanu, Decebal Constantin [3 ]
Nguyen, Phuong H. [1 ]
Liotta, Antonio [4 ]
Webber, Michael E. [5 ]
Gibescu, Madeleine [1 ]
Slootweg, J. G. [1 ]
机构
[1] Eindhoven Univ Technol, Dept Elect Engn, NL-5600 MB Eindhoven, Netherlands
[2] Eindhoven Univ Technol, Dept Mech Engn, NL-5600 MB Eindhoven, Netherlands
[3] Eindhoven Univ Technol, Dept Math & Comp Sci, NL-5600 MB Eindhoven, Netherlands
[4] Univ Derby, Data Sci Ctr, Derby DE1 3HD, England
[5] Univ Texas Austin, Dept Mech Engn, Austin, TX 78712 USA
基金
欧盟地平线“2020”;
关键词
Deep reinforcement learning; demand response; deep neural networks; smart grid; strategic optimization; PREDICTION;
D O I
10.1109/TSG.2018.2834219
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Unprecedented high volumes of data are becoming available with the growth of the advanced metering infrastructure. These arc expected to benefit planning and operation of the future power systems and to help customers transition from a passive to an active role. In this paper, we explore for the first time in the smart grid context the benefits of using deep reinforcement learning, a hybrid type of methods that combines reinforcement learning with deep learning, to perform on-line optimization of schedules for building energy management systems. The learning procedure was explored using two methods, Deep Q-learning and deep policy gradient, both of which have been extended to perform multiple actions simultaneously. The proposed approach was validated on the large-scale Pecan Street Inc. database. This highly dimensional database includes information about photovoltaic power generation, electric vehicles and buildings appliances. Moreover, these on-line energy scheduling strategies could be used to provide real-time feedback to consumers to encourage more efficient use of electricity.
引用
收藏
页码:3698 / 3708
页数:11
相关论文
empty
未找到相关数据