Asset correlation based deep reinforcement learning for the portfolio selection

被引:17
作者
Zhao, Tianlong [1 ]
Ma, Xiang [1 ]
Li, Xuemei [1 ]
Zhang, Caiming [1 ,2 ,3 ]
机构
[1] Shandong Univ, Sch Software, Jinan 250101, Peoples R China
[2] Shandong Coinnovat Ctr Future Intelligent Comp, Yantai 264025, Peoples R China
[3] Digital Media Technol Key Lab Shandong Prov, Jinan 250014, Peoples R China
基金
中国国家自然科学基金;
关键词
Portfolio selection; Deep reinforcement learning; Asset correlations; PREDICTION; REVERSION;
D O I
10.1016/j.eswa.2023.119707
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Portfolio selection is an important application of AI in the financial field, which has attracted considerable attention from academia and industry alike. One of the great challenges in this application is modeling the correlation among assets in the portfolio. However, current studies cannot deal well with this challenge because it is difficult to analyze complex nonlinearity in the correlation. This paper proposes a policy network that models the nonlinear correlation by utilizing the self-attention mechanism to better tackle this issue. In addition, a deterministic policy gradient recurrent reinforcement learning method based on Monte Carlo sampling is constructed with the objective function of cumulative return to train the policy network. In most existing reinforcement learning-based studies, the state transition probability is generally regarded as unknown, so the value function of the policy can only be estimated. Based on financial backtest experiments, we analyze that the state transition probability is known in the portfolio, and value function can be directly obtained by sampling, further theoretically proving the optimality of the proposed reinforcement learning method in the portfolio. Finally, the superiority and generality of our approach are demonstrated through comprehensive experiments on the cryptocurrency dataset, S&P 500 stock dataset, and ETF dataset.
引用
收藏
页数:15
相关论文
共 59 条
  • [1] Agarwal A., 2006, P 23 INT C MACH LEAR, P9, DOI DOI 10.1145/1143844.1143846
  • [2] Bai SJ, 2018, Arxiv, DOI [arXiv:1803.01271, DOI 10.48550/ARXIV.1803.01271]
  • [3] Borodin A., 2003, Advances in Neural Information Processing Systems, V16
  • [4] Multi-DQN: An ensemble of Deep Q-learning agents for stock market forecasting
    Carta, Salvatore
    Ferreira, Anselmo
    Podda, Alessandro Sebastian
    Recupero, Diego Reforgiato
    Sanna, Antonio
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2021, 164
  • [5] Constructing a multilayer network for stock market
    Chen, Wei
    Jiang, Manrui
    Jiang, Cheng
    [J]. SOFT COMPUTING, 2020, 24 (09) : 6345 - 6361
  • [6] Cover T. M., 1991, Math. Financ., V1, P1, DOI DOI 10.1111/J.1467-9965.1991.TB00002.X
  • [7] Das P., 2011, P ACM SIGKDD INT C K, P1163, DOI DOI 10.1145/2020408.2020588
  • [8] Computational learning techniques for intraday FX trading using popular technical indicators
    Dempster, MAH
    Payne, TW
    Romahi, Y
    Thompson, GWP
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2001, 12 (04): : 744 - 754
  • [9] Deep Direct Reinforcement Learning for Financial Signal Representation and Trading
    Deng, Yue
    Bao, Feng
    Kong, Youyong
    Ren, Zhiquan
    Dai, Qionghai
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (03) : 653 - 664
  • [10] Frye J., 2008, The Journal of Credit Risk, V4, P75, DOI DOI 10.21314/JCR.2008.071