Reinforcement Learning Trees

被引:112
|
作者
Zhu, Ruoqing [1 ]
Zeng, Donglin [1 ]
Kosorok, Michael R. [1 ]
机构
[1] Univ N Carolina, Dept Biostat, CB 7420, Chapel Hill, NC 27599 USA
关键词
Consistency; Error bound; Random forests; Reinforcement learning; Trees; RANDOM FORESTS; RANDOMIZED TREES; CLASSIFICATION; SELECTION; CONSISTENCY; LASSO; MODEL;
D O I
10.1080/01621459.2015.1036994
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In this article, we introduce a new type of tree-based method, reinforcement learning trees (RLT), which exhibits significantly improved performance over traditional methods such as random forests (Breiman 2001) under high-dimensional settings. The innovations are threefold. First, the new method implements reinforcement learning at each selection of a splitting variable during the tree construction processes. By splitting on the variable that brings the greatest future improvement in later splits, rather than choosing the one with largest marginal effect from the immediate split, the constructed tree uses the available samples in a more efficient way. Moreover, such an approach enables linear combination cuts at little extra computational cost. Second, we propose a variable muting procedure that progressively eliminates noise variables during the construction of each individual tree. The muting procedure also takes advantage of reinforcement learning and prevents noise variables from being considered in the search for splitting rules, so that toward terminal nodes, where the sample size is small, the splitting rules are still constructed from only strong variables. Last, we investigate asymptotic properties of the proposed method under basic assumptions and discuss rationale in general settings. Supplementary materials for this article are available online.
引用
收藏
页码:1770 / 1784
页数:15
相关论文
共 50 条
  • [1] Adaptive building of decision trees by reinforcement learning
    Preda, Mircea
    PROCEEDINGS OF THE 7TH WSEAS INTERNATIONAL CONFERENCE ON APPLIED INFORMATICS AND COMMUNICATIONS, 2007, : 34 - 39
  • [2] Improving reinforcement learning by using sequence trees
    Sertan Girgin
    Faruk Polat
    Reda Alhajj
    Machine Learning, 2010, 81 : 283 - 331
  • [3] Evolving interpretable decision trees for reinforcement learning
    Costa, Vinicius G.
    Perez-Aracil, Jorge
    Salcedo-Sanz, Sancho
    Pedreira, Carlos E.
    ARTIFICIAL INTELLIGENCE, 2024, 327
  • [4] Improving reinforcement learning by using sequence trees
    Girgin, Sertan
    Polat, Faruk
    Alhajj, Reda
    MACHINE LEARNING, 2010, 81 (03) : 283 - 331
  • [5] Cost-Sensitive Trees for Interpretable Reinforcement Learning
    Nishtala, Siddharth
    Ravindran, Balaraman
    PROCEEDINGS OF 7TH JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MANAGEMENT OF DATA, CODS-COMAD 2024, 2024, : 91 - 99
  • [6] Learning to see the wood for the trees: machine learning, decision trees, and the classification of isolated theropod teeth
    Wills, Simon
    Underwood, Charlie J.
    Barrett, Paul M.
    PALAEONTOLOGY, 2021, 64 (01) : 75 - 99
  • [7] Exploring Continuous Action Spaces with Diffusion Trees for Reinforcement Learning
    Vollmer, Christian
    Schaffernicht, Erik
    Gross, Horst-Michael
    ARTIFICIAL NEURAL NETWORKS-ICANN 2010, PT II, 2010, 6353 : 190 - 199
  • [8] From Reinforcement Learning to Deep Reinforcement Learning: An Overview
    Agostinelli, Forest
    Hocquet, Guillaume
    Singh, Sameer
    Baldi, Pierre
    BRAVERMAN READINGS IN MACHINE LEARNING: KEY IDEAS FROM INCEPTION TO CURRENT STATE, 2018, 11100 : 298 - 328
  • [9] Uncertainty Quantification and Exploration for Reinforcement Learning
    Zhu, Yi
    Dong, Jing
    Lam, Henry
    OPERATIONS RESEARCH, 2024, 72 (04) : 1689 - 1709
  • [10] Recursively Imputed Survival Trees
    Zhu, Ruoqing
    Kosorok, Michael R.
    JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2012, 107 (497) : 331 - 340