Reinforcement learning building control approach harnessing imitation learning

被引:27
作者
Dey, Sourav [1 ]
Marzullo, Thibault [2 ]
Zhang, Xiangyu [2 ]
Henze, Gregor [1 ,2 ,3 ]
机构
[1] Univ Colorado, Dept Civil Environm & Architectural Engn, Boulder, CO 80309 USA
[2] Natl Renewable Energy Lab, Golden, CO USA
[3] Renewable & Sustainable Energy Inst, Boulder, CO 80303 USA
关键词
Reinforcement learning; Building controls; Imitation learning; Artificial intelligence;
D O I
10.1016/j.egyai.2023.100255
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning (RL) has shown significant success in sequential decision making in fields like autonomous vehicles, robotics, marketing and gaming industries. This success has attracted the attention to the RL control approach for building energy systems which are becoming complicated due to the need to optimize for multiple, potentially conflicting, goals like occupant comfort, energy use and grid interactivity. However, for real world applications, RL has several drawbacks like requiring large training data and time, and unstable control behavior during the early exploration process making it infeasible for an application directly to building control tasks. To address these issues, an imitation learning approach is utilized herein where the RL agents starts with a policy transferred from accepted rule based policies and heuristic policies. This approach is successful in reducing the training time, preventing the unstable early exploration behavior and improving upon an accepted rule-based policy - all of these make RL a more practical control approach for real world applications in the domain of building controls.
引用
收藏
页数:13
相关论文
共 50 条
[41]   A Motion Capture and Imitation Learning Based Approach to Robot Control [J].
Racinskis, Peteris ;
Arents, Janis ;
Greitans, Modris .
APPLIED SCIENCES-BASEL, 2022, 12 (14)
[42]   DeFIX: Detecting and Fixing Failure Scenarios with Reinforcement Learning in Imitation Learning Based Autonomous Driving [J].
Dagdanov, Resul ;
Eksen, Feyza ;
Durmus, Halil ;
Yurdakul, Ferhat ;
Ure, Nazim Kemal .
2022 IEEE 25TH INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2022, :4215-4220
[43]   Comparison of reinforcement learning and model predictive control for building energy system optimization [J].
Wang, Dan ;
Zheng, Wanfu ;
Wang, Zhe ;
Wang, Yaran ;
Pang, Xiufeng ;
Wang, Wei .
APPLIED THERMAL ENGINEERING, 2023, 228
[44]   Harnessing deep reinforcement learning algorithms for image categorization: A multi algorithm approach [J].
Yerramreddy, Dhanvanth Reddy ;
Marasani, Jayasurya ;
Ponnuru, Sathwik Venkata Gowtham ;
Min, Dugki ;
Don, S. .
ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 136
[45]   Decentralized multi-region perimeter control in complex urban environments using reinforcement and imitation learning [J].
Kampitakis, Emmanouil ;
Vlahogianni, Eleni I. .
TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2025, 178
[46]   An integrated imitation and reinforcement learning methodology for robust agile aircraft control with limited pilot demonstration data [J].
Sever, Gulay Goktas ;
Demir, Umut ;
Satir, A. Sadik ;
Sahin, Mustafa Cagatay ;
Ure, Nazim Kemal .
AEROSPACE SCIENCE AND TECHNOLOGY, 2025, 158
[47]   Reinforcement learning for building controls: The opportunities and challenges [J].
Wang, Zhe ;
Hong, Tianzhen .
APPLIED ENERGY, 2020, 269
[48]   The reinforcement learning method for occupant behavior in building control: A review [J].
Han M. ;
Zhao J. ;
Zhang X. ;
Shen J. ;
Li Y. .
Energy and Built Environment, 2021, 2 (02) :137-148
[49]   Adaptive Control for Building Energy Management Using Reinforcement Learning [J].
Eller, Lukas ;
Siafara, Lydia C. ;
Sauter, Thilo .
2018 IEEE INTERNATIONAL CONFERENCE ON INDUSTRIAL TECHNOLOGY (ICIT), 2018, :1562-1567
[50]   A Simulation Study on Energy Optimization in Building Control with Reinforcement Learning [J].
Bolt, Peter ;
Ziebart, Volker ;
Jaeger, Christian ;
Schmid, Nicolas ;
Stadelmann, Thilo ;
Fuechslin, Rudolf M. .
ARTIFICIAL NEURAL NETWORKS IN PATTERN RECOGNITION, ANNPR 2024, 2024, 15154 :320-331