Statistical arbitrage trading on the intraday market using the asynchronous advantage actor-critic method

被引:10
|
作者
Demir, Sumeyra
Stappers, Bart
Kok, Koen
Paterakis, Nikolaos G.
机构
[1] Department of Electrical Engineering, Eindhoven University of Technology
[2] Scholt Energy, Valkensvard
关键词
Algorithmic trading; Actor-critic; A3C; Behaviour cloning; Deep reinforcement learning; Intraday markets; Non-physical virtual trader; Single intraday coupled market; Statistical arbitrage; ELECTRICITY; IMPACT;
D O I
10.1016/j.apenergy.2022.118912
中图分类号
TE [石油、天然气工业]; TK [能源与动力工程];
学科分类号
0807 ; 0820 ;
摘要
In this paper, we focus on statistical arbitrage trading opportunities involving the continuous exploitation of price differences arising during an intraday trading period with the option of closing positions on the balancing market. We aim to maximise the reward-risk ratio of an autonomous trading strategy. To find an optimal trading policy, we propose utilising the asynchronous advantage actor-critic (A3C) algorithm, a deep reinforcement learning method, with function approximators of two-headed shared deep neural networks. We enforce a risk-constrained trading strategy by limiting the maximum allowed position, and conduct state engineering and selection processes. We introduce a novel reward function and goal-based exploration, i.e. behaviour cloning. Our methodology is evaluated on a case study using the limit order book of the European single intraday coupled market (SIDC) available for the Dutch market area. The majority of hourly products on the test set return a profit. We expect our study to benefit electricity traders, renewable electricity producers and researchers who seek to implement state-of-art intelligent trading strategies.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Adaptive bias-variance trade-off in advantage estimator for actor-critic algorithms
    Chen, Yurou
    Zhang, Fengyi
    Liu, Zhiyong
    NEURAL NETWORKS, 2024, 169 : 764 - 777
  • [32] Balance Control for the First-order Inverted Pendulum Based on the Advantage Actor-critic Algorithm
    Yan Zheng
    Xutong Li
    Long Xu
    International Journal of Control, Automation and Systems, 2020, 18 : 3093 - 3100
  • [33] Cost-effective dynamic sampling in high dimensional online monitoring with advantage actor-critic
    Li, Haoqian
    Chen, Yudong
    Liu, Kaibo
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2025,
  • [34] On the sample complexity of actor-critic method for reinforcement learning with function approximation
    Kumar, Harshat
    Koppel, Alec
    Ribeiro, Alejandro
    MACHINE LEARNING, 2023, 112 (07) : 2433 - 2467
  • [35] Balance Control for the First-order Inverted Pendulum Based on the Advantage Actor-critic Algorithm
    Zheng, Yan
    Li, Xutong
    Xu, Long
    INTERNATIONAL JOURNAL OF CONTROL AUTOMATION AND SYSTEMS, 2020, 18 (12) : 3093 - 3100
  • [36] On the sample complexity of actor-critic method for reinforcement learning with function approximation
    Harshat Kumar
    Alec Koppel
    Alejandro Ribeiro
    Machine Learning, 2023, 112 : 2433 - 2467
  • [37] Large-scale Interactive Conversational Recommendation System using Actor-Critic Framework
    Montazeralghaem, Ali
    Allan, James
    Thomas, Philip S.
    15TH ACM CONFERENCE ON RECOMMENDER SYSTEMS (RECSYS 2021), 2021, : 220 - 229
  • [38] DAG-based workflows scheduling using Actor-Critic Deep Reinforcement Learning
    Koslovski, Guilherme Piegas
    Pereira, Kleiton
    Albuquerque, Paulo Roberto
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 150 : 354 - 363
  • [39] Optimal Tracking Control for Robotic Manipulator using Actor-Critic Network
    Hu, Yong
    Cui, Lingguo
    Chai, Senchun
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 1556 - 1561
  • [40] Fault Diagnosis for Gas Turbine Rotor Using Actor-Critic Network
    Cui, Yingjie
    Wang, Hongjun
    PROCEEDINGS OF TEPEN 2022, 2023, 129 : 923 - 935