Classification and regression trees offer straightforward methods of attributing importance values to input features, either globally or for a single prediction. Conditional feature contributions (CFCs) yield local, case-by-case explanations of a prediction by following the decision path and attributing changes in the expected output of the model to each feature along the path. However, CFCs suffer from a potential bias which depends on the distance from the root of a tree. The by now immensely popular alternative, SHapley Additive exPlanation (SHAP) values appear to mitigate this bias but are computationally much more expensive. Here we contribute a thorough, empirical comparison of the explanations computed by both methods on a set of 164 publicly available classification problems in order to provide data-driven algorithm recommendations to current researchers. For random forests and boosted trees, we find extremely high similarities and correlations of both local and global SHAP values and CFC scores, leading to very similar rankings and interpretations. Unsurprisingly, these insights extend to the fidelity of using global feature importance scores as a proxy for the predictive power associated with each feature.