Supervised actor-critic reinforcement learning with action feedback for algorithmic trading

被引:5
|
作者
Sun, Qizhou [1 ]
Si, Yain-Whar [1 ]
机构
[1] Univ Macau, Dept Comp & Informat Sci, Ave da Univ, Taipa, Macau, Peoples R China
关键词
Finance; Reinforcement learning; Supervised learning; Algorithmic trading; ENERGY;
D O I
10.1007/s10489-022-04322-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is one of the promising approaches for algorithmic trading in financial markets. However, in certain situations, buy or sell orders issued by an algorithmic trading program may not be fulfilled entirely. By considering the actual scenarios from the financial markets, in this paper, we propose a novel framework named Supervised Actor-Critic Reinforcement Learning with Action Feedback (SACRL-AF) for solving this problem. The action feedback mechanism of SACRL-AF notifies the actor about the dealt positions and corrects the transitions of the replay buffer. Meanwhile, the dealt positions are used as the labels for the supervised learning. Recent studies have shown that Deep Deterministic Policy Gradient (DDPG) and Twin Delayed Deep Deterministic Policy Gradient (TD3) are more stable and superior to other actor-critic algorithms. Against this background, based on the proposed SACRL-AF framework, two reinforcement learning algorithms henceforth referred to as Supervised Deep Deterministic Policy Gradient with Action Feedback (SDDPG-AF) and Supervised Twin Delayed Deep Deterministic Policy Gradient with Action Feedback (STD3-AF) are proposed in this paper. Experimental results show that SDDPG-AF and STD3-AF achieve the state-of-art performance in profitability.
引用
收藏
页码:16875 / 16892
页数:18
相关论文
共 50 条
  • [41] Reinforcement learning for automatic quadrilateral mesh generation: A soft actor-critic approach
    Pan, Jie
    Huang, Jingwei
    Cheng, Gengdong
    Zeng, Yong
    NEURAL NETWORKS, 2023, 157 : 288 - 304
  • [42] Dynamic Content Caching Based on Actor-Critic Reinforcement Learning for IoT Systems
    Lai, Lifeng
    Zheng, Fu-Chun
    Wen, Wanli
    Luo, Jingjing
    Li, Ge
    2022 IEEE 96TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2022-FALL), 2022,
  • [43] A bounded actor-critic reinforcement learning algorithm applied to airline revenue management
    Lawhead, Ryan J.
    Gosavi, Abhijit
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2019, 82 : 252 - 262
  • [44] Combining backpropagation with Equilibrium Propagation to improve an Actor-Critic reinforcement learning framework
    Kubo, Yoshimasa
    Chalmers, Eric
    Luczak, Artur
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2022, 16
  • [45] Dynamic spectrum access and sharing through actor-critic deep reinforcement learning
    Dong, Liang
    Qian, Yuchen
    Xing, Yuan
    EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING, 2022, 2022 (01)
  • [46] Intensive versus non-intensive actor-critic reinforcement learning algorithms
    Wawrzynski, P
    Pacut, A
    ARTIFICIAL INTELLIGENCE AND SOFT COMPUTING - ICAISC 2004, 2004, 3070 : 934 - 941
  • [47] SMONAC: Supervised Multiobjective Negative Actor-Critic for Sequential Recommendation
    Zhou, Fei
    Luo, Biao
    Wu, Zhengke
    Huang, Tingwen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 13
  • [48] Locating algorithm of steel stock area with asynchronous advantage actor-critic reinforcement learning
    Cho, Young-in
    Kim, Byeongseop
    Yoon, Hee-Chang
    Woo, Jong Hun
    JOURNAL OF COMPUTATIONAL DESIGN AND ENGINEERING, 2024, 11 (01) : 230 - 246
  • [49] Maximizing Information Usefulness in Vehicular CP Networks Using Actor-Critic Reinforcement Learning
    Ghnaya, Imed
    Ahmed, Toufik
    Mosbah, Mohamed
    Aniss, Hasnaa
    2022 18TH INTERNATIONAL CONFERENCE ON NETWORK AND SERVICE MANAGEMENT (CNSM 2022): INTELLIGENT MANAGEMENT OF DISRUPTIVE NETWORK TECHNOLOGIES AND SERVICES, 2022, : 296 - 302
  • [50] A Continuous Actor-Critic Reinforcement Learning Approach to Flocking with Fixed-Wing UAVs
    Wang, Chang
    Yan, Chao
    Xiang, Xiaojia
    Zhou, Han
    ASIAN CONFERENCE ON MACHINE LEARNING, VOL 101, 2019, 101 : 64 - 79