Soft Actor-Critic Deep Reinforcement Learning with Hybrid Mixed-Integer Actions for Demand Responsive Scheduling of Energy Systems

被引:11
|
作者
Campos, Gustavo [1 ]
El-Farra, Nael H. [1 ]
Palazoglu, Ahmet [1 ]
机构
[1] Univ Calif Davis, Dept Chem Engn, Davis, CA 95616 USA
关键词
ALGORITHMS; MODEL; GAME; GO;
D O I
10.1021/acs.iecr.1c04984
中图分类号
TQ [化学工业];
学科分类号
0817 ;
摘要
We investigate the application of a Deep Reinforcement Learning (DRL) method for demand responsive closed-loop scheduling of continuous process/energy systems. The method employed is the Soft Actor-Critic (SAC), an actor-critic, off-policy, stochastic method with built-in entropy maximization that balances exploration and exploitation. Considering energy systems, which are typically characterized by the presence of hybrid (combined discrete-continuous) actions originating from equipment operating ranges and discrete actuators, we demonstrate the main ways in which hybrid actions can be incorporated into the SAC framework. A unified treatment is presented, in which five different approaches for modeling hybrid actions are compared: two considering deterministic discrete decisions (DSReL and Softmax), and three considering stochastic discrete decisions (Q-enumeration, Gumbel-Softmax reparameterization, and Score Function gradient estimator). It is shown that DSReL and Q-enumeration have a better overall performance for the considered environment. Next, the developed hybrid-SAC method is applied to the operation of process/energy systems under day-ahead electricity prices and demand forecast. A case-study of a large-scale District Cooling plant employing real demand and price data is presented. It is shown that the algorithm can quickly avoid constraint violations, and continuously improves toward the optimal solution. Lastly, an analysis of demand forecast uncertainty shows that the hybrid-SAC algorithm can robustly handle state uncertainty and works well for partially observable systems with incomplete state information.
引用
收藏
页码:8443 / 8461
页数:19
相关论文
共 50 条
  • [1] Soft Actor-Critic With Integer Actions
    Fan, Ting-Han
    Wang, Yubo
    2022 AMERICAN CONTROL CONFERENCE, ACC, 2022, : 2611 - 2616
  • [2] Averaged Soft Actor-Critic for Deep Reinforcement Learning
    Ding, Feng
    Ma, Guanfeng
    Chen, Zhikui
    Gao, Jing
    Li, Peng
    COMPLEXITY, 2021, 2021
  • [3] Integrated Actor-Critic for Deep Reinforcement Learning
    Zheng, Jiaohao
    Kurt, Mehmet Necip
    Wang, Xiaodong
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 505 - 518
  • [4] An Actor-Critic Deep Reinforcement Learning Approach for Transmission Scheduling in Cognitive Internet of Things Systems
    Yang, Helin
    Xie, Xianzhong
    IEEE SYSTEMS JOURNAL, 2020, 14 (01): : 51 - 60
  • [5] Actor-Critic Deep Reinforcement Learning for Solving Job Shop Scheduling Problems
    Liu, Chien-Liang
    Chang, Chuan-Chin
    Tseng, Chun-Jan
    IEEE ACCESS, 2020, 8 : 71752 - 71762
  • [6] Visual Navigation with Actor-Critic Deep Reinforcement Learning
    Shao, Kun
    Zhao, Dongbin
    Zhu, Yuanheng
    Zhang, Qichao
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [7] Deep Actor-Critic Reinforcement Learning for Anomaly Detection
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [8] An Actor-Critic Reinforcement Learning Approach for Energy Harvesting Communications Systems
    Masadeh, Ala'eddin
    Wang, Zhengdao
    Kamal, Ahmed E.
    2019 28TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND NETWORKS (ICCCN), 2019,
  • [9] User Scheduling and Resource Allocation in HetNets With Hybrid Energy Supply: An Actor-Critic Reinforcement Learning Approach
    Wei, Yifei
    Yu, F. Richard
    Song, Mei
    Han, Zhu
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2018, 17 (01) : 680 - 692
  • [10] On-line Energy Optimization of Hybrid Production Systems Using Actor-Critic Reinforcement Learning
    Schwung, Dorothea
    Schwung, Andreas
    Ding, Steven X.
    2018 9TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS (IS), 2018, : 147 - 154