共 1 条
Asynchronous Deep Double Dueling Q-learning for trading-signal execution in limit order book markets
被引:0
|作者:
Nagy, Peer
[1
]
Calliess, Jan-Peter
[1
]
Zohren, Stefan
[1
,2
,3
]
机构:
[1] Univ Oxford, Oxford Man Inst Quantitat Finance, Dept Engn Sci, Oxford OX1 2JD, England
[2] Man Grp, London, England
[3] Alan Turing Inst, London, England
来源:
FRONTIERS IN ARTIFICIAL INTELLIGENCE
|
2023年
/
6卷
关键词:
limit order books;
quantitative finance;
reinforcement learning;
LOBSTER;
algorithmic trading;
D O I:
10.3389/frai.2023.1151003
中图分类号:
TP18 [人工智能理论];
学科分类号:
081104 ;
0812 ;
0835 ;
1405 ;
摘要:
We employ deep reinforcement learning (RL) to train an agent to successfully translate a high-frequency trading signal into a trading strategy that places individual limit orders. Based on the ABIDES limit order book simulator, we build a reinforcement learning OpenAI gym environment and utilize it to simulate a realistic trading environment for NASDAQ equities based on historic order book messages. To train a trading agent that learns to maximize its trading return in this environment, we use Deep Dueling Double Q-learning with the APEX (asynchronous prioritized experience replay) architecture. The agent observes the current limit order book state, its recent history, and a short-term directional forecast. To investigate the performance of RL for adaptive trading independently from a concrete forecasting algorithm, we study the performance of our approach utilizing synthetic alpha signals obtained by perturbing forward-looking returns with varying levels of noise. Here, we find that the RL agent learns an effective trading strategy for inventory management and order placing that outperforms a heuristic benchmark trading strategy having access to the same signal.
引用
收藏
页数:10
相关论文