Statistical arbitrage trading on the intraday market using the asynchronous advantage actor-critic method

被引:10
|
作者
Demir, Sumeyra
Stappers, Bart
Kok, Koen
Paterakis, Nikolaos G.
机构
[1] Department of Electrical Engineering, Eindhoven University of Technology
[2] Scholt Energy, Valkensvard
关键词
Algorithmic trading; Actor-critic; A3C; Behaviour cloning; Deep reinforcement learning; Intraday markets; Non-physical virtual trader; Single intraday coupled market; Statistical arbitrage; ELECTRICITY; IMPACT;
D O I
10.1016/j.apenergy.2022.118912
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
In this paper, we focus on statistical arbitrage trading opportunities involving the continuous exploitation of price differences arising during an intraday trading period with the option of closing positions on the balancing market. We aim to maximise the reward-risk ratio of an autonomous trading strategy. To find an optimal trading policy, we propose utilising the asynchronous advantage actor-critic (A3C) algorithm, a deep reinforcement learning method, with function approximators of two-headed shared deep neural networks. We enforce a risk-constrained trading strategy by limiting the maximum allowed position, and conduct state engineering and selection processes. We introduce a novel reward function and goal-based exploration, i.e. behaviour cloning. Our methodology is evaluated on a case study using the limit order book of the European single intraday coupled market (SIDC) available for the Dutch market area. The majority of hourly products on the test set return a profit. We expect our study to benefit electricity traders, renewable electricity producers and researchers who seek to implement state-of-art intelligent trading strategies.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Merging with Extraction Method for Transfer Learning in Actor-Critic
    Takano, Toshiaki
    Takase, Haruhiko
    Kawanaka, Hiroharu
    Tsuruoka, Shinji
    JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2011, 15 (07) : 814 - 821
  • [22] Learning-Based Resource Allocation in Cloud Data Center Using Advantage Actor-Critic
    Chen, Zheyi
    Hu, Jia
    Min, Geyong
    ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [23] An Actor-Critic Method for Simulation-Based Optimization
    Li, Kuo
    Jia, Qing-Shan
    Yan, Jiaqi
    IFAC PAPERSONLINE, 2022, 55 (11): : 7 - 12
  • [24] An Improved Actor-Critic Method for AutoCombating In Infantry Vehicles
    Chen, Ruizhu
    Fei, Rong
    Wu, Lili
    Yang, Lu
    2022 IEEE INTERNATIONAL CONFERENCE ON TRUST, SECURITY AND PRIVACY IN COMPUTING AND COMMUNICATIONS, TRUSTCOM, 2022, : 1162 - 1167
  • [25] Asynchronous learning for actor-critic neural networks and synchronous triggering for multiplayer system
    Wang, Ke
    Mu, Chaoxu
    ISA TRANSACTIONS, 2022, 129 : 295 - 308
  • [26] Dynamic User Resource Allocation for Downlink Multicarrier NOMA with an Actor-Critic Method
    Wang, Xinshui
    Meng, Ke
    Wang, Xu
    Liu, Zhibin
    Ma, Yuefeng
    ENERGIES, 2023, 16 (07)
  • [27] ACIS: An Improved Actor-Critic Method for POMDPs with Internal State
    Xu, Dan
    Liu, Quan
    2015 IEEE 27TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2015), 2015, : 369 - 376
  • [28] Extractive text summarization model based on advantage actor-critic and graph matrix methodology
    Yang, Senqi
    Duan, Xuliang
    Wang, Xi
    Tang, Dezhao
    Xiao, Zeyan
    Guo, Yan
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (01) : 1488 - 1504
  • [29] DASH Live Video Streaming Control Using Actor-Critic Reinforcement Learning Method
    Wei, Bo
    Song, Hang
    Quang Ngoc Nguyen
    Katto, Jiro
    MOBILE NETWORKS AND MANAGEMENT, MONAMI 2021, 2022, 418 : 17 - 24
  • [30] Enhancing Autonomous Driving Navigation Using Soft Actor-Critic
    Elallid, Badr Ben
    Benamar, Nabil
    Bagaa, Miloud
    Hadjadj-Aoul, Yassine
    FUTURE INTERNET, 2024, 16 (07)