Multiadvisor Reinforcement Learning for Multiagent Multiobjective Smart Home Energy Control

被引:11
作者
Tittaferrante A. [1 ]
Yassine A. [2 ]
机构
[1] Lakehead University, Department of Electrical and Computer Engineering, Thunder Bay, P7B5E1, ON
[2] Lakehead University, Software Engineering Department, Thunder Bay, P7B5E1, ON
来源
IEEE Transactions on Artificial Intelligence | 2022年 / 3卷 / 04期
关键词
Deep learning; energy consumption; multiagent; multiobjective; reinforcement learning; smart home;
D O I
10.1109/TAI.2021.3125918
中图分类号
学科分类号
摘要
Effective automated smart home energy control is essential for smart grid approaches to demand response (DR). This is a multiobjective adaptive control problem because it balances an appliance's primary objective with demand response objectives. One challenge comes from the heterogeneous nature of objectives, requiring tradeoffs between comfort, cost, and other objectives. Another challenge comes from the heterogeneous dynamics, which result from different environments and the different appliances used. Another challenge is nonstationary nature of dynamics and rewards due to seasonal changes and time-varying user preferences. Finally, we consider computational challenges, required by the real-time aspect of the control problem, particularly notable due to 'the curse of dimensionality.' We propose a multiagent multiadvisor reinforcement learning framework to address these challenges. We design a smart-home simulation to demonstrate the performance (in terms of weighted reward) of our approach relative to competitive single-objective reinforcement learning algorithms. Furthermore, we theoretically and empirically demonstrate the linear computational scalability of the algorithm. Finally, we identify the need for key performance measures of the proposed system by considering the effect of selected preferences on agents. Overall, the proposed algorithm is reasonably competitive with conventional approaches while simultaneously enabling behavior changes with change in preferences without requiring more data. © 2020 IEEE.
引用
收藏
页码:581 / 594
页数:13
相关论文
共 42 条
  • [1] Singh S., Yassine A., Mining energy consumption behavior patterns for households in smart grid, IEEE Trans. Emerg. Topics Comput., 7, 3, pp. 404-419, (2019)
  • [2] Samad T., Koch E., Stluka P., Automated demand response for smart buildings and microgrids: The state of the practice and research challenges, Proc. IEEE IRE, 104, 4, pp. 726-744, (2016)
  • [3] Sutton R.S., Barto A.G., Reinforcement Learning: An Introduction, (2018)
  • [4] Mnih V., Et al., Human-level control through deep reinforcement learning, Nature, 518, 7540, pp. 529-533, (2015)
  • [5] Laroche R., Fatemi M., Romoff J., Van Seijen H., Multi-advisor reinforcement learning, (2017)
  • [6] Wan Z., Li H., He H., Prokhorov D., Model-free real-time EV charging scheduling based on deep reinforcement learning, IEEE Trans. Smart Grid, 10, 5, pp. 5246-5257, (2019)
  • [7] O'Neill D., Levorato M., Goldsmith A., Mitra U., Residential demand response using reinforcement learning, Proc. 1st IEEE Int. Conf. Smart Grid Commun., pp. 409-414, (2010)
  • [8] Watkins C.J., Dayan P., Technical note: Q-learning, Mach. Learn., 8, 3, pp. 279-292, (1992)
  • [9] Wen Z., O'Neill D., Maei H., Optimal demand response using devicebased reinforcement learning, IEEE Trans. Smart Grid, 6, 5, pp. 2312-2324, (2015)
  • [10] Althaher S., Mancarella P., Mutale J., Automated demand response from home energy management system under dynamic pricing and power and comfort constraints, IEEE Trans. Smart Grid, 6, 4, pp. 1874-1883, (2015)