Deep Reinforcement Learning for Dynamic Stock Option Hedging: A Review

被引:2
|
作者
Pickard, Reilly [1 ]
Lawryshyn, Yuri [2 ]
机构
[1] Univ Toronto, Dept Mech & Ind Engn, Toronto, ON M5S 3G8, Canada
[2] Univ Toronto, Dept Chem Engn & Appl Chem, Toronto, ON M5S 3E5, Canada
关键词
reinforcement learning; neural networks; dynamic stock option hedging; quantitative finance; financial risk management; VOLATILITY;
D O I
10.3390/math11244943
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
This paper reviews 17 studies addressing dynamic option hedging in frictional markets through Deep Reinforcement Learning (DRL). Specifically, this work analyzes the DRL models, state and action spaces, reward formulations, data generation processes and results for each study. It is found that policy methods such as DDPG are more commonly employed due to their suitability for continuous action spaces. Despite diverse state space definitions, a lack of consensus exists on variable inclusion, prompting a call for thorough sensitivity analyses. Mean-variance metrics prevail in reward formulations, with episodic return, VaR and CvaR also yielding comparable results. Geometric Brownian motion is the primary data generation process, supplemented by stochastic volatility models like SABR (stochastic alpha, beta, rho) and the Heston model. RL agents, particularly those monitoring transaction costs, consistently outperform the Black-Scholes Delta method in frictional environments. Although consistent results emerge under constant and stochastic volatility scenarios, variations arise when employing real data. The lack of a standardized testing dataset or universal benchmark in the RL hedging space makes it difficult to compare results across different studies. A recommended future direction for this work is an implementation of DRL for hedging American options and an investigation of how DRL performs compared to other numerical American option hedging methods.
引用
收藏
页数:19
相关论文
共 50 条
  • [1] Application of deep learning to option hedging strategy
    Li, Haibin
    Sun, Mingzhu
    Wang, Ziqi
    SYSTEMS AND SOFT COMPUTING, 2024, 6
  • [2] Structured products dynamic hedging based on reinforcement learning
    Xu H.
    Xu C.
    Yan H.
    Sun Y.
    Journal of Ambient Intelligence and Humanized Computing, 2023, 14 (09) : 12285 - 12295
  • [3] Deep-Reinforcement-Learning-Based Dynamic Ensemble Model for Stock Prediction
    Lin, Wenjing
    Xie, Liang
    Xu, Haijiao
    ELECTRONICS, 2023, 12 (21)
  • [4] A deep structure for option discovery in Reinforcement Learning
    Mohammadi, Jahanbakhsh
    Mozayani, Nasser
    2016 SMART GRIDS CONFERENCE (SGC), 2016, : 65 - 68
  • [5] Deep reinforcement learning in chemistry: A review
    Sridharan, Bhuvanesh
    Sinha, Animesh
    Bardhan, Jai
    Modee, Rohit
    Ehara, Masahiro
    Priyakumar, U. Deva
    JOURNAL OF COMPUTATIONAL CHEMISTRY, 2024, 45 (22) : 1886 - 1898
  • [6] Preliminary remarks on option pricing and dynamic hedging
    Fliess, Michel
    Join, Cedric
    2012 1ST INTERNATIONAL CONFERENCE ON SYSTEMS AND COMPUTER SCIENCE (ICSCS), 2012,
  • [7] Gamma and vega hedging using deep distributional reinforcement learning
    Cao, Jay
    Chen, Jacky
    Farghadani, Soroush
    Hull, John
    Poulos, Zissis
    Wang, Zeyu
    Yuan, Jun
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2023, 6
  • [8] Deep Hedging: Continuous Reinforcement Learning for Hedging of General Portfolios across Multiple Risk Aversions
    Murray, Phillip
    Wood, Ben
    Buehler, Hans
    Wiese, Magnus
    Pakkanen, Mikko S.
    3RD ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2022, 2022, : 361 - 368
  • [9] Application of deep reinforcement learning in stock trading strategies and stock forecasting
    Yuming Li
    Pin Ni
    Victor Chang
    Computing, 2020, 102 : 1305 - 1322
  • [10] Application of deep reinforcement learning in stock trading strategies and stock forecasting
    Li, Yuming
    Ni, Pin
    Chang, Victor
    COMPUTING, 2020, 102 (06) : 1305 - 1322