Supervised actor-critic reinforcement learning with action feedback for algorithmic trading

被引:5
|
作者
Sun, Qizhou [1 ]
Si, Yain-Whar [1 ]
机构
[1] Univ Macau, Dept Comp & Informat Sci, Ave da Univ, Taipa, Macau, Peoples R China
关键词
Finance; Reinforcement learning; Supervised learning; Algorithmic trading; ENERGY;
D O I
10.1007/s10489-022-04322-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is one of the promising approaches for algorithmic trading in financial markets. However, in certain situations, buy or sell orders issued by an algorithmic trading program may not be fulfilled entirely. By considering the actual scenarios from the financial markets, in this paper, we propose a novel framework named Supervised Actor-Critic Reinforcement Learning with Action Feedback (SACRL-AF) for solving this problem. The action feedback mechanism of SACRL-AF notifies the actor about the dealt positions and corrects the transitions of the replay buffer. Meanwhile, the dealt positions are used as the labels for the supervised learning. Recent studies have shown that Deep Deterministic Policy Gradient (DDPG) and Twin Delayed Deep Deterministic Policy Gradient (TD3) are more stable and superior to other actor-critic algorithms. Against this background, based on the proposed SACRL-AF framework, two reinforcement learning algorithms henceforth referred to as Supervised Deep Deterministic Policy Gradient with Action Feedback (SDDPG-AF) and Supervised Twin Delayed Deep Deterministic Policy Gradient with Action Feedback (STD3-AF) are proposed in this paper. Experimental results show that SDDPG-AF and STD3-AF achieve the state-of-art performance in profitability.
引用
收藏
页码:16875 / 16892
页数:18
相关论文
共 50 条
  • [31] Dynamic spectrum access and sharing through actor-critic deep reinforcement learning
    Liang Dong
    Yuchen Qian
    Yuan Xing
    EURASIP Journal on Wireless Communications and Networking, 2022
  • [32] SOFT ACTOR-CRITIC REINFORCEMENT LEARNING FOR ROBOTIC MANIPULATOR WITH HINDSIGHT EXPERIENCE REPLAY
    Yan, Tao
    Zhang, Wenan
    Yang, Simon X.
    Yu, Li
    INTERNATIONAL JOURNAL OF ROBOTICS & AUTOMATION, 2019, 34 (05): : 536 - 543
  • [33] MULTI-STEP ACTOR-CRITIC FRAMEWORK FOR REINFORCEMENT LEARNING IN CONTINUOUS CONTROL
    Huang T.
    Chen G.
    Journal of Applied and Numerical Optimization, 2023, 5 (02): : 189 - 200
  • [34] Actor-Critic for Multi-Agent Reinforcement Learning with Self-Attention
    Zhao, Juan
    Zhu, Tong
    Xiao, Shuo
    Gao, Zongqian
    Sun, Hao
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (09)
  • [35] SAC-FACT: Soft Actor-Critic Reinforcement Learning for Counterfactual Explanations
    Ezzeddine, Fatima
    Ayoub, Omran
    Andreoletti, Davide
    Giordano, Silvia
    EXPLAINABLE ARTIFICIAL INTELLIGENCE, XAI 2023, PT I, 2023, 1901 : 195 - 216
  • [36] IMPROVING ACTOR-CRITIC REINFORCEMENT LEARNING VIA HAMILTONIAN MONTE CARLO METHOD
    Xu, Duo
    Fekri, Faramarz
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4018 - 4022
  • [37] CONTROLLED SENSING AND ANOMALY DETECTION VIA SOFT ACTOR-CRITIC REINFORCEMENT LEARNING
    Zhong, Chen
    Gursoy, M. Cenk
    Velipasalar, Senem
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4198 - 4202
  • [38] Anomaly Detection Under Controlled Sensing Using Actor-Critic Reinforcement Learning
    Joseph, Geethu
    Gursoy, M. Cenk
    Varshney, Pramod K.
    PROCEEDINGS OF THE 21ST IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (IEEE SPAWC2020), 2020,
  • [39] Automatic collective motion tuning using actor-critic deep reinforcement learning
    Abpeikar, Shadi
    Kasmarik, Kathryn
    Garratt, Matthew
    Hunjet, Robert
    Khan, Md Mohiuddin
    Qiu, Huanneng
    SWARM AND EVOLUTIONARY COMPUTATION, 2022, 72
  • [40] An Actor-critic Reinforcement Learning Model for Optimal Bidding in Online Display Advertising
    Yuan, Congde
    Guo, Mengzhuo
    Xiang, Chaoneng
    Wang, Shuangyang
    Song, Guoqing
    Zhang, Qingpeng
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 3604 - 3613