Optimal Policy Characterization Enhanced Actor-Critic Approach for Electric Vehicle Charging Scheduling in a Power Distribution Network

被引:71
作者
Jin, Jiangliang [1 ,2 ]
Xu, Yunjian [1 ]
机构
[1] Chinese Univ Hong Kong, Dept Mech & Automat Engn, Hong Kong, Peoples R China
[2] Chinese Univ Hong Kong, Shun Hing Inst Adv Engn, Hong Kong, Peoples R China
关键词
Electric vehicle charging; Optimal scheduling; Stochastic processes; Distribution networks; Reinforcement learning; Solar power generation; Dynamic programming; deep reinforcement learning; electric vehicle charging; actor-critic approach; power distribution network; DEMAND RESPONSE; ENERGY-STORAGE; REINFORCEMENT; SYSTEMS;
D O I
10.1109/TSG.2020.3028470
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We study the scheduling of large-scale electric vehicle (EV) charging in a power distribution network under random renewable generation and electricity prices. The problem is formulated as a stochastic dynamic program with unknown state transition probability. To mitigate the curse of dimensionality, we establish the nodal multi-target (NMT) characterization of the optimal scheduling policy: all EVs with the same deadline at the same bus should be charged to approach a single target of remaining energy demand. We prove that the NMT characterization is optimal under arbitrarily random system dynamics. To adaptively learn the dynamics of system uncertainty, we propose a model-free soft-actor-critic (SAC) based method to determine the target levels for the characterized NMT policy. The proposed SAC + NMT approach significantly outperforms existing deep reinforcement learning methods (in our numerical experiments on the IEEE 37-node test feeder), as the established NMT characterization sharply reduces the dimensionality of neural network outputs without loss of optimality.
引用
收藏
页码:1416 / 1428
页数:13
相关论文
共 47 条
  • [11] Chen SY, 2012, IEEE POW ENER SOC GE
  • [12] Reinforcement Learning-Based Plug-in Electric Vehicle Charging With Forecasted Price
    Chis, Adriana
    Lunden, Jarmo
    Koivunen, Visa
    [J]. IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2017, 66 (05) : 3674 - 3684
  • [13] Coordination of Electric Vehicle Charging Through Multiagent Reinforcement Learning
    Da Silva, Felipe Leno
    Nishida, Cyntia E. H.
    Roijers, Diederik M.
    Costa, Anna H. Reali
    [J]. IEEE TRANSACTIONS ON SMART GRID, 2020, 11 (03) : 2347 - 2356
  • [14] Farivar M, 2013, IEEE DECIS CONTR P, P4329, DOI 10.1109/CDC.2013.6760555
  • [15] Haarnoja T., 2018, SOFT ACTOR CRITIC AL
  • [16] Haarnoja T, 2018, PR MACH LEARN RES, V80
  • [17] Hill A., 2018, Stable baselines
  • [18] MULTILAYER FEEDFORWARD NETWORKS ARE UNIVERSAL APPROXIMATORS
    HORNIK, K
    STINCHCOMBE, M
    WHITE, H
    [J]. NEURAL NETWORKS, 1989, 2 (05) : 359 - 366
  • [19] Matching EV Charging Load With Uncertain Wind: A Simulation-Based Policy Improvement Approach
    Huang, Qilong
    Jia, Qing-Shan
    Qiu, Zhifeng
    Guan, Xiaohong
    Deconinck, Geert
    [J]. IEEE TRANSACTIONS ON SMART GRID, 2015, 6 (03) : 1425 - 1433
  • [20] Konda VR, 2000, ADV NEUR IN, V12, P1008