Back to Search
Start Over
Large Scale Prediction with Decision Trees.
- Source :
-
Journal of the American Statistical Association . Mar2024, Vol. 119 Issue 545, p525-537. 13p. - Publication Year :
- 2024
-
Abstract
- This article shows that decision trees constructed with Classification and Regression Trees (CART) and C4.5 methodology are consistent for regression and classification tasks, even when the number of predictor variables grows sub-exponentially with the sample size, under natural 0-norm and 1-norm sparsity constraints. The theory applies to a wide range of models, including (ordinary or logistic) additive regression models with component functions that are continuous, of bounded variation, or, more generally, Borel measurable. Consistency holds for arbitrary joint distributions of the predictor variables, thereby accommodating continuous, discrete, and/or dependent data. Finally, we show that these qualitative properties of individual trees are inherited by Breiman's random forests. A key step in the analysis is the establishment of an oracle inequality, which allows for a precise characterization of the goodness of fit and complexity tradeoff for a mis-specified model. for this article are available online. [ABSTRACT FROM AUTHOR]
Details
- Language :
- English
- ISSN :
- 01621459
- Volume :
- 119
- Issue :
- 545
- Database :
- Academic Search Index
- Journal :
- Journal of the American Statistical Association
- Publication Type :
- Academic Journal
- Accession number :
- 175846048
- Full Text :
- https://doi.org/10.1080/01621459.2022.2126782