Back to Search Start Over

Large Scale Prediction with Decision Trees.

Authors :
Klusowski, Jason M.
Tian, Peter M.
Source :
Journal of the American Statistical Association. Mar2024, Vol. 119 Issue 545, p525-537. 13p.
Publication Year :
2024

Abstract

This article shows that decision trees constructed with Classification and Regression Trees (CART) and C4.5 methodology are consistent for regression and classification tasks, even when the number of predictor variables grows sub-exponentially with the sample size, under natural 0-norm and 1-norm sparsity constraints. The theory applies to a wide range of models, including (ordinary or logistic) additive regression models with component functions that are continuous, of bounded variation, or, more generally, Borel measurable. Consistency holds for arbitrary joint distributions of the predictor variables, thereby accommodating continuous, discrete, and/or dependent data. Finally, we show that these qualitative properties of individual trees are inherited by Breiman's random forests. A key step in the analysis is the establishment of an oracle inequality, which allows for a precise characterization of the goodness of fit and complexity tradeoff for a mis-specified model. for this article are available online. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
01621459
Volume :
119
Issue :
545
Database :
Academic Search Index
Journal :
Journal of the American Statistical Association
Publication Type :
Academic Journal
Accession number :
175846048
Full Text :
https://doi.org/10.1080/01621459.2022.2126782