Reinforcement Learning Trees

被引:112
|
作者
Zhu, Ruoqing [1 ]
Zeng, Donglin [1 ]
Kosorok, Michael R. [1 ]
机构
[1] Univ N Carolina, Dept Biostat, CB 7420, Chapel Hill, NC 27599 USA
关键词
Consistency; Error bound; Random forests; Reinforcement learning; Trees; RANDOM FORESTS; RANDOMIZED TREES; CLASSIFICATION; SELECTION; CONSISTENCY; LASSO; MODEL;
D O I
10.1080/01621459.2015.1036994
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In this article, we introduce a new type of tree-based method, reinforcement learning trees (RLT), which exhibits significantly improved performance over traditional methods such as random forests (Breiman 2001) under high-dimensional settings. The innovations are threefold. First, the new method implements reinforcement learning at each selection of a splitting variable during the tree construction processes. By splitting on the variable that brings the greatest future improvement in later splits, rather than choosing the one with largest marginal effect from the immediate split, the constructed tree uses the available samples in a more efficient way. Moreover, such an approach enables linear combination cuts at little extra computational cost. Second, we propose a variable muting procedure that progressively eliminates noise variables during the construction of each individual tree. The muting procedure also takes advantage of reinforcement learning and prevents noise variables from being considered in the search for splitting rules, so that toward terminal nodes, where the sample size is small, the splitting rules are still constructed from only strong variables. Last, we investigate asymptotic properties of the proposed method under basic assumptions and discuss rationale in general settings. Supplementary materials for this article are available online.
引用
收藏
页码:1770 / 1784
页数:15
相关论文
共 50 条
  • [41] Reinforcement learning algorithms: A brief survey
    Shakya, Ashish Kumar
    Pillai, Gopinatha
    Chakrabarty, Sohom
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 231
  • [42] From Reinforcement Learning to Knowledge of Nature
    Red'ko, V. G.
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2023, 33 (03) : 478 - 482
  • [43] Reinforcement learning for ridesharing: An extended survey
    Qin, Zhiwei
    Zhu, Hongtu
    Ye, Jieping
    TRANSPORTATION RESEARCH PART C-EMERGING TECHNOLOGIES, 2022, 144
  • [44] Continuous-time reinforcement learning approach for portfolio management with time penalization
    Garcia-Galicia, Mauricio
    Carsteanu, Alin A.
    Clempner, Julio B.
    EXPERT SYSTEMS WITH APPLICATIONS, 2019, 129 : 27 - 36
  • [45] Characterizing reinforcement learning methods through parameterized learning problems
    Kalyanakrishnan, Shivaram
    Stone, Peter
    MACHINE LEARNING, 2011, 84 (1-2) : 205 - 247
  • [46] Adaptive humanoid robot behaviour in a serious game scenario through reinforcement learning
    Zedda, Eleonora
    Manca, Marco
    Paterno, Fabio
    Santoro, Carmen
    BEHAVIOUR & INFORMATION TECHNOLOGY, 2025,
  • [47] GREY SYSTEM BASED REACTIVE NAVIGATION OF MOBILE ROBOTS USING REINFORCEMENT LEARNING
    Chen, Chunlin
    Dong, Daoyi
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2010, 6 (02): : 789 - 800
  • [48] Reinforcement and learning
    Maria R. Servedio
    Stein A. Sæther
    Glenn-Peter Sætre
    Evolutionary Ecology, 2009, 23 : 109 - 123
  • [49] Reinforcement and learning
    Servedio, Maria R.
    Saether, Stein A.
    Saetre, Glenn-Peter
    EVOLUTIONARY ECOLOGY, 2009, 23 (01) : 109 - 123
  • [50] Towards Automated Imbalanced Learning with Deep Hierarchical Reinforcement Learning
    Zha, Daochen
    Lai, Kwei-Herng
    Tan, Qiaoyu
    Ding, Sirui
    Zou, Na
    Hu, Xia Ben
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2476 - 2485