Deep Reinforcement Learning for Dynamic Stock Option Hedging: A Review

被引:2
|
作者
Pickard, Reilly [1 ]
Lawryshyn, Yuri [2 ]
机构
[1] Univ Toronto, Dept Mech & Ind Engn, Toronto, ON M5S 3G8, Canada
[2] Univ Toronto, Dept Chem Engn & Appl Chem, Toronto, ON M5S 3E5, Canada
关键词
reinforcement learning; neural networks; dynamic stock option hedging; quantitative finance; financial risk management; VOLATILITY;
D O I
10.3390/math11244943
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
This paper reviews 17 studies addressing dynamic option hedging in frictional markets through Deep Reinforcement Learning (DRL). Specifically, this work analyzes the DRL models, state and action spaces, reward formulations, data generation processes and results for each study. It is found that policy methods such as DDPG are more commonly employed due to their suitability for continuous action spaces. Despite diverse state space definitions, a lack of consensus exists on variable inclusion, prompting a call for thorough sensitivity analyses. Mean-variance metrics prevail in reward formulations, with episodic return, VaR and CvaR also yielding comparable results. Geometric Brownian motion is the primary data generation process, supplemented by stochastic volatility models like SABR (stochastic alpha, beta, rho) and the Heston model. RL agents, particularly those monitoring transaction costs, consistently outperform the Black-Scholes Delta method in frictional environments. Although consistent results emerge under constant and stochastic volatility scenarios, variations arise when employing real data. The lack of a standardized testing dataset or universal benchmark in the RL hedging space makes it difficult to compare results across different studies. A recommended future direction for this work is an implementation of DRL for hedging American options and an investigation of how DRL performs compared to other numerical American option hedging methods.
引用
收藏
页数:19
相关论文
共 50 条
  • [21] Dynamic Network Slicing using Deep Reinforcement Learning
    Kumar, Swaraj
    Vankayala, Satya Kumar
    Singh, Devashish
    Roy, Ishaan
    Sahoo, Biswa P. S.
    Yoon, Seungil
    Kanakaraj, Ignatius Samuel
    2021 IEEE INTERNATIONAL CONFERENCE ON ADVANCED NETWORKS AND TELECOMMUNICATIONS SYSTEMS (IEEE ANTS), 2021,
  • [22] Caching in Dynamic IoT Networks by Deep Reinforcement Learning
    Yao, Jingjing
    Ansari, Nirwan
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (05) : 3268 - 3275
  • [23] Geometric deep reinforcement learning for dynamic DAG scheduling
    Grinsztajn, Nathan
    Beaumont, Olivier
    Jeannot, Emmanuel
    Preux, Philippe
    2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 258 - 265
  • [24] Reinforcement Learning in Dynamic Task Scheduling: A Review
    Shyalika C.
    Silva T.
    Karunananda A.
    SN Computer Science, 2020, 1 (6)
  • [25] Dynamic Multitarget Assignment Based on Deep Reinforcement Learning
    Wu, Yifei
    Lei, Yonglin
    Zhu, Zhi
    Yang, Xiaochen
    Li, Qun
    IEEE ACCESS, 2022, 10 : 75998 - 76007
  • [26] Adaptive stock trading with dynamic asset allocation using reinforcement learning
    O, Jangmin
    Lee, Jongwoo
    Lee, Jae Won
    Zhang, Byoung-Tak
    INFORMATION SCIENCES, 2006, 176 (15) : 2121 - 2147
  • [27] HEDGING BARRIER OPTIONS USING REINFORCEMENT LEARNING
    Chen, Jacky
    Fu, Yu
    Hull, John
    Poulos, Zissis
    Wang, Zeyu
    Yuan, Jun
    JOURNAL OF INVESTMENT MANAGEMENT, 2024, 22 (04): : 16 - 25
  • [28] Deep Reinforcement Learning: An Overview
    Mousavi, Seyed Sajad
    Schukat, Michael
    Howley, Enda
    PROCEEDINGS OF SAI INTELLIGENT SYSTEMS CONFERENCE (INTELLISYS) 2016, VOL 2, 2018, 16 : 426 - 440
  • [29] Reinforcement Learning in Stock Trading
    Quang-Vinh Dang
    ADVANCED COMPUTATIONAL METHODS FOR KNOWLEDGE ENGINEERING (ICCSAMA 2019), 2020, 1121 : 311 - 322
  • [30] Deep learning and reinforcement learning approach on microgrid
    Chandrasekaran, Kumar
    Kandasamy, Prabaakaran
    Ramanathan, Srividhya
    INTERNATIONAL TRANSACTIONS ON ELECTRICAL ENERGY SYSTEMS, 2020, 30 (10):