A Deep Residual Shrinkage Neural Network-based Deep Reinforcement Learning Strategy in Financial Portfolio Management

被引:13
作者
Sun, Ruoyu [1 ]
Jiang, Zhengyong [1 ]
Su, Jionglong [2 ]
机构
[1] Xian Jiaotong Liverpool Univ, Dept Math Sci, Suzhou, Peoples R China
[2] Xian Jiaotong Liverpool Univ, XJTLU Entrepreneur Coll Taicang, Sch AI & Adv Comp, Suzhou, Peoples R China
来源
2021 IEEE 6TH INTERNATIONAL CONFERENCE ON BIG DATA ANALYTICS (ICBDA 2021) | 2021年
关键词
deep reinforcement learning; residual network; residual shrinkage network; cryptocurrency; algorithmic trading; portfolio management;
D O I
10.1109/ICBDA51983.2021.9403210
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement Learning algorithms are widely applied in many fields, such as price index prediction, image recognition, and natural language processing. This paper introduces a novel algorithm based on the classical Deep Reinforcement Learning algorithm and Deep Residual Shrinkage Neural Network for portfolio management. In this algorithm, the Ensemble of Identical Independent Evaluators framework put forward by Jiang et al. is adopted in the policy function. Following this, we adopt the Deep Residual Shrinkage Neural Network to function as the identical independent evaluator to optimize the algorithm. We use the cryptocurrency market in this research to assess the efficacy of our strategy with eight traditional portfolio management strategies as well as Jiang et al.'s reinforcement learning strategy. In our experiments, the Accumulated Yield is used to reflect the profit of the algorithm. Despite having a high commission rate of 0.25% in back-tests, results show that our algorithm can achieve 44.5% 105.4% and 148.8% returns in three different 50-days back-tests, which is five times more than the profit of other non-reinforcement learning strategies and Jiang et al.'s strategy. Furthermore, the Sharpe ratio demonstrates that the extra reward per unit risk of the our strategy is still better than other traditional portfolio management strategies and Jiang et al.'s strategy by at least 50% in different time horizons.
引用
收藏
页码:76 / 86
页数:11
相关论文
共 32 条
[1]  
Agarwal A., 2006, P 23 INT C MACH LEAR, P9
[2]  
[Anonymous], 2014, arXiv
[3]  
[Anonymous], 2016, CONTINUOUS CONTROL D, DOI DOI 10.1561/2200000006
[4]   Can we learn to beat the best stock [J].
Borodin, A ;
El-Yaniv, R ;
Gogan, V .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2004, 21 :579-594
[5]  
Cover T.M., 1991, Mathematical finance, V1, P1, DOI 10.1111/j.1467-9965.1991.tb00002.x
[6]   Deep Direct Reinforcement Learning for Financial Signal Representation and Trading [J].
Deng, Yue ;
Bao, Feng ;
Kong, Youyong ;
Ren, Zhiquan ;
Dai, Qionghai .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (03) :653-664
[7]  
Funt Mark J, 2010, Pa Dent J (Harrisb), V77, P33
[8]   A Framework of Hierarchical Deep Q-Network for Portfolio Management [J].
Gao, Yuan ;
Gao, Ziming ;
Hu, Yi ;
Song, Sifan ;
Jiang, Zhengyong ;
Su, Jionglong .
ICAART: PROCEEDINGS OF THE 13TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE - VOL 2, 2021, :132-140
[9]  
Gao ZM, 2020, 2020 5TH IEEE INTERNATIONAL CONFERENCE ON BIG DATA ANALYTICS (IEEE ICBDA 2020), P268, DOI 10.1109/ICBDA49040.2020.9101333
[10]   Identity Mappings in Deep Residual Networks [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
COMPUTER VISION - ECCV 2016, PT IV, 2016, 9908 :630-645