Autonomous HVAC Control, A Reinforcement Learning Approach

被引:87
作者
Barrett, Enda [1 ,2 ]
Linder, Stephen [1 ,2 ]
机构
[1] Schneider Elect, Galway, Ireland
[2] Schneider Elect, Andover, MA 01810 USA
来源
MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, PT III | 2015年 / 9286卷
关键词
HVAC control; Reinforcement learning; Bayesian learning;
D O I
10.1007/978-3-319-23461-8_1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent high profile developments of autonomous learning thermostats by companies such as Nest Labs and Honeywell have brought to the fore the possibility of ever greater numbers of intelligent devices permeating our homes and working environments into the future. However, the specific learning approaches and methodologies utilised by these devices have never been made public. In fact little information is known as to the specifics of how these devices operate and learn about their environments or the users who use them. This paper proposes a suitable learning architecture for such an intelligent thermostat in the hope that it will benefit further investigation by the research community. Our architecture comprises a number of different learning methods each of which contributes to create a complete autonomous thermostat capable of controlling a HVAC system. A novel state action space formalism is proposed to enable a Reinforcement Learning agent to successfully control the HVAC system by optimising both occupant comfort and energy costs. Our results show that the learning thermostat can achieve cost savings of 10% over a programmable thermostat, whilst maintaining high occupant comfort standards.
引用
收藏
页码:3 / 19
页数:17
相关论文
共 21 条
[1]  
Ahmed O., 2004, METHOD APPARATUS DET, Patent No. 2,289,237
[2]  
[Anonymous], 2000, P 17 INT C MACH LEAR
[3]  
[Anonymous], 1995, ARTIFICIAL INTELLIGE
[4]  
[Anonymous], 2004, AUTOMATED PLANNING T
[5]  
Barrett E., 2011, 2011 IEEE 9th European Conference on Web Services, P83, DOI 10.1109/ECOWS.2011.27
[6]  
Barrett E, 2012, CONCURRENCY COMPUTAT
[7]   A parallel framework for Bayesian reinforcement learning [J].
Barrett, Enda ;
Duggan, Jim ;
Howley, Enda .
CONNECTION SCIENCE, 2014, 26 (01) :7-23
[8]  
Choi SPM, 1996, ADV NEUR IN, V8, P945
[9]  
Dage G., 1998, METHOD SYSTEM CONTRO
[10]  
Dorigo M., 2014, Proceedings of ML-95, P252