1. Computational Efficieucy in the Selection of Regression Variables
- Author
-
L. R. LaMotte and R. R. Hocking
- Subjects
Statistics and Probability ,Mathematical optimization ,Variables ,Proper linear model ,Applied Mathematics ,media_common.quotation_subject ,Explained sum of squares ,Regression analysis ,Residual sum of squares ,Modeling and Simulation ,Linear predictor function ,Statistics ,Segmented regression ,Regression diagnostic ,media_common ,Mathematics - Abstract
A number of criteria have been proposed for selecting the best subset or subsets of independent variables in linear regression analysis. Applying these criteria to all possible subsets is, in general, not feasible if the number of variables is large. Many of the criteria are monotone functions of the residual sum of squares hence the problem is reduced to identifying subsets for which this quantity is small. In an earlier paper (Selection of the Best Subset in Regression Analysis by R. R. Hocking and R. N. Leslie, 1967) a method was described for identifying such subsets without considering all possible subsets. However, the amount of computation required if more than fifteen independent variables were considered was excessive. The present paper extends the basic ideas in that paper so that moderately large problems can how be treated with what appears to be a minimum of computation.
- Published
- 1970