Reinforcement learning for the optimization of electric vehicle virtual power plants

被引:17
作者
Al-Gabalawy, Mostafa [1 ]
机构
[1] Future Univ Egypt, Fac Comp & Informat Technol, Dept Digital Media Technol, Cairo, Egypt
关键词
distributed energy resources; electric vehicle; machine learning; optimization; reinforcement learning; virtual power plants; BIDDING STRATEGY; MARKETS; INTEGRATION; ENERGY; FLEETS; MODEL;
D O I
10.1002/2050-7038.12951
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Integrating weather-dependent renewable energy sources into the electricity system impose challenges on the power grid. Balancing services are needed, which can be provided by virtual power plants (VPP) that aggregate distributed energy resources (DER) to consume or produce electricity on demand. Electric vehicle (EV) fleets can use idle cars' batteries as combined storage to offer balancing services on smart electricity markets. However, there are risks associated with this business model extension. The fleet faces severe imbalance penalties if it cannot charge the offered amount of balancing energy due to the vehicles' unpredicted mobility demand. Ensuring the fleet can fulfill all market commitments risks denying profitable customer rentals. We study the design of a decision support system that estimates these risks, dynamically adjusts the composition of a VPP portfolio, and profitably places bids on multiple electricity markets simultaneously. Here we show that a reinforcement learning agent can optimize the VPP portfolio by learning from favorable market conditions and fleet demand uncertainties. In comparison to previous research, in which the bidding risks were unknown and fleets could only offer conservative amounts of balancing power to a single market, our proposed approach increases the amount of offered balancing power by 48% to 82% and achieves a charging cost reduction of the fleet by 25%. In experiments with real-world carsharing data of 500 EVs, we found that mobility demand forecasting algorithms' accuracy is crucial for a successful bidding strategy. Moreover, we show that recent advancements in deep reinforcement learning decrease the convergence time and improve the results' robustness. Our results demonstrate how modern RL algorithms can be successfully used for fleet management, VPP optimization, and demand response in the smart grid. We anticipate that DER, such as EVs, will play an essential role in providing reliable backup power for the grid and formulate market design recommendations to allow easier access to these resources.
引用
收藏
页数:30
相关论文
共 65 条
[1]  
Abadi M, 2016, ACM SIGPLAN NOTICES, V51, P1, DOI [10.1145/3022670.2976746, 10.1145/2951913.2976746]
[2]  
Agricola A., 2014, DENA ANCILLARY SERVI
[3]  
Al-Gabalawy Mostafa, 2021, ISA Trans, DOI 10.1016/j.isatra.2021.01.005
[4]   RETRACTED: Optimal peak shifting of a domestic load connected to utility grid using storage battery based on deepQ-learningnetwork (Retracted Article) [J].
Al-Gablawy, Mostafa .
INTERNATIONAL JOURNAL OF ENERGY RESEARCH, 2021, 45 (02) :3269-3287
[5]  
[Anonymous], 2011, AAAI
[6]   Managing electricity price modeling risk via ensemble forecasting: The case of Turkey [J].
Avci, Ezgi ;
Ketter, Wolfgang ;
van Heck, Eric .
ENERGY POLICY, 2018, 123 :390-403
[7]   NEURONLIKE ADAPTIVE ELEMENTS THAT CAN SOLVE DIFFICULT LEARNING CONTROL-PROBLEMS [J].
BARTO, AG ;
SUTTON, RS ;
ANDERSON, CW .
IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS, 1983, 13 (05) :834-846
[8]   DYNAMIC PROGRAMMING [J].
BELLMAN, R .
SCIENCE, 1966, 153 (3731) :34-&
[9]   Designing Smart Markets [J].
Bichler, Martin ;
Gupta, Alok ;
Ketter, Wolfgang .
INFORMATION SYSTEMS RESEARCH, 2010, 21 (04) :688-699
[10]  
BMU, 2010, EN CONC ENV SOUND RE