Modeling limit order trading with a continuous action policy for deep reinforcement learning

被引:4
|
作者
Tsantekidis, Avraam [1 ]
Passalis, Nikolaos [1 ]
Tefas, Anastasios [1 ]
机构
[1] Aristotle Univ Thessaloniki, Sch Informat, Thessaloniki, Greece
关键词
Financial trading; Limit orders; Policy gradient; Deep reinforcement learning; PREDICTION; MARKET;
D O I
10.1016/j.neunet.2023.05.051
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Limit Orders allow buyers and sellers to set a "limit price"they are willing to accept in a trade. On the other hand, market orders allow for immediate execution at any price. Thus, market orders are susceptible to slippage, which is the additional cost incurred due to the unfavorable execution of a trade order. As a result, limit orders are often preferred, since they protect traders from excessive slippage costs due to larger than expected price fluctuations. Despite the price guarantees of limit orders, they are more complex compared to market orders. Orders with overly optimistic limit prices might never be executed, which increases the risk of employing limit orders in Machine Learning (ML)-based trading systems. Indeed, the current ML literature for trading almost exclusively relies on market orders. To overcome this limitation, a Deep Reinforcement Learning (DRL) approach is proposed to model trading agents that use limit orders. The proposed method (a) uses a framework that employs a continuous probability distribution to model limit prices, while (b) provides the ability to place market orders when the risk of no execution is more significant than the cost of slippage. Extensive experiments are conducted with multiple currency pairs, using hourly price intervals, validating the effectiveness of the proposed method and paving the way for introducing limit order modeling in DRL-based trading.& COPY; 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:506 / 515
页数:10
相关论文
共 50 条
  • [41] Deep Reinforcement Learning for Quantitative Trading: Challenges and Opportunities
    An, Bo
    Sun, Shuo
    Wang, Rundong
    IEEE INTELLIGENT SYSTEMS, 2022, 37 (02) : 23 - 26
  • [42] Deep Reinforcement Learning for Trading-A Critical Survey
    Millea, Adrian
    DATA, 2021, 6 (11)
  • [43] Policy Reuse in Deep Reinforcement Learning
    Glatt, Ruben
    Helena, Anna
    Costa, Reali
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4929 - 4930
  • [44] Deep learning for limit order books
    Sirignano, Justin A.
    QUANTITATIVE FINANCE, 2019, 19 (04) : 549 - 570
  • [45] A Comparison of Action Selection Methods for Implicit Policy Method Reinforcement Learning in Continuous Action-Space
    Nichols, Barry D.
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 3785 - 3792
  • [46] Action-specialized expert ensemble trading system with extended discrete action space using deep reinforcement learning
    Leem, JoonBum
    Kim, Ha Young
    PLOS ONE, 2020, 15 (07):
  • [47] Policy Derivation Methods for Critic-Only Reinforcement Learning in Continuous Action Spaces
    Alibekov, Eduard
    Kubalik, Jiri
    Babuska, Robert
    IFAC PAPERSONLINE, 2016, 49 (05): : 285 - 290
  • [48] Discretizing Continuous Action Space With Unimodal Probability Distributions for On-Policy Reinforcement Learning
    Zhu, Yuanyang
    Wang, Zhi
    Zhu, Yuanheng
    Chen, Chunlin
    Zhao, Dongbin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [49] Explicit Bandwidth Learning for FOREX Trading Using Deep Reinforcement Learning
    Nalmpantis, Angelos
    Passalis, Nikolaos
    Tefas, Anastasios
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 686 - 690
  • [50] Goal-Oriented Obstacle Avoidance with Deep Reinforcement Learning in Continuous Action Space
    Cimurs, Reinis
    Lee, Jin Han
    Suh, Il Hong
    ELECTRONICS, 2020, 9 (03)