Enhancing Grid-Interactive Buildings Demand Response: Sequential Update-Based Multiagent Deep Reinforcement Learning Approach

被引:5
作者
Han, Yinghua [1 ,2 ]
Wu, Jingrun [1 ]
Chen, Haoqi [1 ]
Si, Fangyuan [3 ,4 ]
Cao, Zhiao [5 ]
Zhao, Qiang [5 ]
机构
[1] Northeastern Univ Qinhuangdao, Sch Comp & Commun Engn, Qinhuangdao 066004, Hebei, Peoples R China
[2] Northeastern Univ Qinhuangdao, Hebei Key Lab Marine Percept Network & Data Proc, Qinhuangdao 066004, Hebei, Peoples R China
[3] Beijing Jiaotong Univ, Sch Elect Engn, Beijing 100044, Peoples R China
[4] Beijing Jiaotong Univ, Beijing Engn Res Ctr Elect Rail Transportat, Beijing 100044, Peoples R China
[5] Northeastern Univ Qinhuangdao, Sch Control Engn, Qinhuangdao 066004, Peoples R China
基金
中国国家自然科学基金;
关键词
Buildings; Energy management; Costs; Demand response; Behavioral sciences; Thermal loading; Computational modeling; Demand response (DR); grid-interactive buildings; multiagent deep reinforcement learning (MADRL); sequential update scheme;
D O I
10.1109/JIOT.2024.3357109
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Demand response (DR) programs aim to improve grid reliability, reduce customer costs, and manage building energy demand. However, coordinating autonomous DR in grid-interactive buildings faces challenges due to uncertainties in system model parameters, complex energy demands, and interdependent operational constraints. Specifically, the energy management problem in grid-interactive buildings is formulated as a partially observable Markov decision process (POMDP), where each building has only local information and no knowledge of other buildings' behavior. This makes achieving consistent coordinated energy management challenging. To address these issues, this article proposes a novel approach called sequential update-based multiagent deep reinforcement learning (SUMADRL). This approach updates the policy of each building sequentially, considering the updates already made by previous buildings. This ensures that the joint policy update of buildings is always highly aligned with the overall energy optimization goal and improves coordinated energy management among buildings. Additionally, the sequential update scheme provides theoretical guarantees for achieving monotonically improving rewards and converging to Nash equilibrium, enhancing credibility and stability. Experimental results show that the proposed method significantly reduces energy costs, achieves efficient load shaping, and demonstrates superior effectiveness and stability in DR compared to the multiagent deep deterministic policy gradient (MADDPG) algorithm.
引用
收藏
页码:24439 / 24451
页数:13
相关论文
共 43 条
[1]   Multiagent Reinforcement Learning for Energy Management in Residential Buildings [J].
Ahrarinouri, Mehdi ;
Rastegar, Mohammad ;
Seifi, Ali Reza .
IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2021, 17 (01) :659-666
[2]  
Ausubel L., 1993, ECON THEOR, V3, P99, DOI [DOI 10.1007/BF01213694, 10.1007/BF01213694]
[3]  
Azar AT, 2015, STUD FUZZ SOFT COMP, V319, P1, DOI 10.1007/978-3-319-12883-2_1
[4]  
Azevedo I., 2013, Rep. THINK_Topic_11
[5]   Reinforcement learning for whole-building HVAC control and demand response [J].
Azuatalam, Donald ;
Lee, Wee-Lih ;
de Nijs, Frits ;
Liebman, Ariel .
ENERGY AND AI, 2020, 2
[6]  
Bedi G., 2018, Ph.D. dissertation
[7]   Deep reinforcement learning to optimise indoor temperature control and heating energy consumption in buildings [J].
Brandi, Silvio ;
Piscitelli, Marco Savino ;
Martellacci, Marco ;
Capozzoli, Alfonso .
ENERGY AND BUILDINGS, 2020, 224
[8]   Model-predictive control and reinforcement learning in multi-energy system case studies [J].
Ceusters, Glenn ;
Rodriguez, Roman Cantu ;
Garcia, Alberte Bouso ;
Franke, Rudiger ;
Deconinck, Geert ;
Helsen, Lieve ;
Nowe, Ann ;
Messagie, Maarten ;
Camargo, Luis Ramirez .
APPLIED ENERGY, 2021, 303
[9]  
Chang YH, 2004, ADV NEUR IN, V16, P807
[10]  
Charbonnier F, 2023, Arxiv, DOI arXiv:2305.18875