Reinforcement Learning Trees

被引:112
|
作者
Zhu, Ruoqing [1 ]
Zeng, Donglin [1 ]
Kosorok, Michael R. [1 ]
机构
[1] Univ N Carolina, Dept Biostat, CB 7420, Chapel Hill, NC 27599 USA
关键词
Consistency; Error bound; Random forests; Reinforcement learning; Trees; RANDOM FORESTS; RANDOMIZED TREES; CLASSIFICATION; SELECTION; CONSISTENCY; LASSO; MODEL;
D O I
10.1080/01621459.2015.1036994
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In this article, we introduce a new type of tree-based method, reinforcement learning trees (RLT), which exhibits significantly improved performance over traditional methods such as random forests (Breiman 2001) under high-dimensional settings. The innovations are threefold. First, the new method implements reinforcement learning at each selection of a splitting variable during the tree construction processes. By splitting on the variable that brings the greatest future improvement in later splits, rather than choosing the one with largest marginal effect from the immediate split, the constructed tree uses the available samples in a more efficient way. Moreover, such an approach enables linear combination cuts at little extra computational cost. Second, we propose a variable muting procedure that progressively eliminates noise variables during the construction of each individual tree. The muting procedure also takes advantage of reinforcement learning and prevents noise variables from being considered in the search for splitting rules, so that toward terminal nodes, where the sample size is small, the splitting rules are still constructed from only strong variables. Last, we investigate asymptotic properties of the proposed method under basic assumptions and discuss rationale in general settings. Supplementary materials for this article are available online.
引用
收藏
页码:1770 / 1784
页数:15
相关论文
共 50 条
  • [31] Category learning in a recurrent neural network with reinforcement learning
    Zhang, Ying
    Pan, Xiaochuan
    Wang, Yihong
    FRONTIERS IN PSYCHIATRY, 2022, 13
  • [32] A reinforcement learning approach to personalized learning recommendation systems
    Tang, Xueying
    Chen, Yunxiao
    Li, Xiaoou
    Liu, Jingchen
    Ying, Zhiliang
    BRITISH JOURNAL OF MATHEMATICAL & STATISTICAL PSYCHOLOGY, 2019, 72 (01) : 108 - 135
  • [33] A Reinforcement Learning Approach to Directed Test Generation for Shared Memory Verification
    Pfeifer, Nicolas
    Zimpel, Bruno, V
    Andrade, Gabriel A. G.
    dos Santos, Luiz C., V
    PROCEEDINGS OF THE 2020 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2020), 2020, : 538 - 543
  • [34] Generation of multiagent animation for object transportation using deep reinforcement learning and blend-trees
    Chen, Shao-Chieh
    Liu, Guan-Ting
    Wong, Sai-Keung
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2021, 32 (3-4)
  • [35] Symbol Generation and Feature Selection for Reinforcement Learning Agents Using Affordances and U-Trees
    Oladell, Marcus
    Huber, Manfred
    PROCEEDINGS 2012 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2012, : 657 - 662
  • [36] FLOW SHOP SCHEDULING WITH REINFORCEMENT LEARNING
    Zhang, Zhicong
    Wang, Weiping
    Zhong, Shouyan
    Hu, Kaishun
    ASIA-PACIFIC JOURNAL OF OPERATIONAL RESEARCH, 2013, 30 (05)
  • [37] Happiness as an intrinsic motivator in reinforcement learning
    Gao, Yue
    Edelman, Shimon
    ADAPTIVE BEHAVIOR, 2016, 24 (05) : 292 - 305
  • [38] Relational Verification using Reinforcement Learning
    Chen, Jia
    Wei, Jiayi
    Feng, Yu
    Bastani, Osbert
    Dillig, Isil
    PROCEEDINGS OF THE ACM ON PROGRAMMING LANGUAGES-PACMPL, 2019, 3 (OOPSLA):
  • [39] Deep reinforcement learning in chemistry: A review
    Sridharan, Bhuvanesh
    Sinha, Animesh
    Bardhan, Jai
    Modee, Rohit
    Ehara, Masahiro
    Priyakumar, U. Deva
    JOURNAL OF COMPUTATIONAL CHEMISTRY, 2024, 45 (22) : 1886 - 1898
  • [40] Reinforcement Learning for Disassembly Task Control
    Weerasekara, Sachini
    Li, Wei
    Isaacs, Jacqueline
    Kamarthi, Sagar
    COMPUTERS & INDUSTRIAL ENGINEERING, 2024, 190