Deep Reinforcement Learning for Dynamic Stock Option Hedging: A Review

被引:2
|
作者
Pickard, Reilly [1 ]
Lawryshyn, Yuri [2 ]
机构
[1] Univ Toronto, Dept Mech & Ind Engn, Toronto, ON M5S 3G8, Canada
[2] Univ Toronto, Dept Chem Engn & Appl Chem, Toronto, ON M5S 3E5, Canada
关键词
reinforcement learning; neural networks; dynamic stock option hedging; quantitative finance; financial risk management; VOLATILITY;
D O I
10.3390/math11244943
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
This paper reviews 17 studies addressing dynamic option hedging in frictional markets through Deep Reinforcement Learning (DRL). Specifically, this work analyzes the DRL models, state and action spaces, reward formulations, data generation processes and results for each study. It is found that policy methods such as DDPG are more commonly employed due to their suitability for continuous action spaces. Despite diverse state space definitions, a lack of consensus exists on variable inclusion, prompting a call for thorough sensitivity analyses. Mean-variance metrics prevail in reward formulations, with episodic return, VaR and CvaR also yielding comparable results. Geometric Brownian motion is the primary data generation process, supplemented by stochastic volatility models like SABR (stochastic alpha, beta, rho) and the Heston model. RL agents, particularly those monitoring transaction costs, consistently outperform the Black-Scholes Delta method in frictional environments. Although consistent results emerge under constant and stochastic volatility scenarios, variations arise when employing real data. The lack of a standardized testing dataset or universal benchmark in the RL hedging space makes it difficult to compare results across different studies. A recommended future direction for this work is an implementation of DRL for hedging American options and an investigation of how DRL performs compared to other numerical American option hedging methods.
引用
收藏
页数:19
相关论文
共 50 条
  • [41] Offline Deep Reinforcement Learning for Dynamic Pricing of Consumer Credit
    Khraishi, Raad
    Okhrati, Ramin
    3RD ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2022, 2022, : 325 - 333
  • [42] Deep Reinforcement Learning Model for Stock Portfolio Management Based on Data Fusion
    Li, Haifeng
    Hai, Mo
    NEURAL PROCESSING LETTERS, 2024, 56 (02)
  • [43] Deep Reinforcement Learning for Dynamic Multichannel Access in Wireless Networks
    Wang, Shangxing
    Liu, Hanpeng
    Gomes, Pedro Henrique
    Krishnamachari, Bhaskar
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2018, 4 (02) : 257 - 265
  • [44] Solving Dynamic Traveling Salesman Problems With Deep Reinforcement Learning
    Zhang, Zizhen
    Liu, Hong
    Zhou, MengChu
    Wang, Jiahai
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (04) : 2119 - 2132
  • [45] Conditionally Elicitable Dynamic Risk Measures for Deep Reinforcement Learning
    Coache, Anthony
    Jaimungal, Sebastian
    Cartea, Alvaro
    SIAM JOURNAL ON FINANCIAL MATHEMATICS, 2023, 14 (04): : 1249 - 1289
  • [46] Dynamic Scheduling in a Flow Shop Using Deep Reinforcement Learning
    Marchesano, Maria Grazia
    Guizzi, Guido
    Santillo, Liberatina Carmela
    Vespoli, Silvestro
    ADVANCES IN PRODUCTION MANAGEMENT SYSTEMS: ARTIFICIAL INTELLIGENCE FOR SUSTAINABLE AND RESILIENT PRODUCTION SYSTEMS, APMS 2021, PT I, 2021, 630 : 152 - 160
  • [47] A Deep Reinforcement Learning-Based Dynamic Computational Offloading Method for Cloud Robotics
    Penmetcha, Manoj
    Min, Byung-Cheol
    IEEE ACCESS, 2021, 9 : 60265 - 60279
  • [48] Deep reinforcement learning for the dynamic and uncertain vehicle routing problem
    Pan, Weixu
    Liu, Shi Qiang
    APPLIED INTELLIGENCE, 2023, 53 (01) : 405 - 422
  • [49] Learning to Navigate Through Complex Dynamic Environment With Modular Deep Reinforcement Learning
    Wang, Yuanda
    He, Haibo
    Sun, Changyin
    IEEE TRANSACTIONS ON GAMES, 2018, 10 (04) : 400 - 412
  • [50] Deep Reinforcement Learning for Multiobjective Optimization
    Li, Kaiwen
    Zhang, Tao
    Wang, Rui
    IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (06) : 3103 - 3114