Towards Better Decision Forests: Forest Alternating Optimization

被引:4
作者
Carreira-Perpinan, Miguel A. [1 ]
Gabidolla, Magzhan [1 ]
Zharmagambetov, Arman [1 ,2 ]
机构
[1] Univ Calif Merced, Dept CSE, Merced, CA 95343 USA
[2] Meta AI FAIR, Menlo Pk, CA USA
来源
2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR | 2023年
关键词
STATISTICAL VIEW;
D O I
10.1109/CVPR52729.2023.00733
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Decision forests are among the most accurate models in machine learning. This is remarkable given that the way they are trained is highly heuristic: neither the individual trees nor the overall forest optimize any well-defined loss. While diversity mechanisms such as bagging or boosting have been until now critical in the success of forests, we think that a better optimization should lead to better forests-ideally eliminating any need for an ensembling heuristic. However, unlike for most other models, such as neural networks, optimizing forests or trees is not easy, because they define a non-differentiable function. We show, for the first time, that it is possible to learn a forest by optimizing a desirable loss and regularization jointly over all its trees and parameters. Our algorithm, Forest Alternating Optimization, is based on defining a forest as a parametric model with a fixed number of trees and structure (rather than adding trees indefinitely as in bagging or boosting). It then iteratively updates each tree in alternation so that the objective function decreases monotonically. The algorithm is so effective at optimizing that it easily overfits, but this can be corrected by averaging. The result is a forest that consistently exceeds the accuracy of the state-of-the-art while using fewer, smaller trees.
引用
收藏
页码:7589 / 7598
页数:10
相关论文
共 37 条
[1]  
[Anonymous], 2018, Advances in Neural Information Processing Systems
[2]  
[Anonymous], 2021, INT J C NEUR NETW IJ
[3]  
Arman Zharmagambetov, 2020, P 37 INT C MACHINE L, P11398
[4]   A random forest guided tour [J].
Biau, Gerard ;
Scornet, Erwan .
TEST, 2016, 25 (02) :197-227
[5]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[6]   Random forests [J].
Breiman, L .
MACHINE LEARNING, 2001, 45 (01) :5-32
[7]  
Breiman L, 1998, ANN STAT, V26, P801
[8]  
Breiman L, 1984, Classification and regression trees, DOI DOI 10.1201/9781315139470
[9]  
Carreira-Perpinan Miguel A., 2020, FODS '20: Proceedings of the 2020 Foundations of Data Science Conference, P35, DOI 10.1145/3412815.3416882
[10]  
Carreira-Perpinan M. A., 2022, TREE ALTERNATING OPT