Abstract
One of the main topic in the development of predictive models is the identification of variables which are predictors of a given outcome. Automated model selection methods, such as backward or forward stepwise regression, are classical solutions to this problem, but are generally based on strong assumptions about the functional form of the model or the distribution of residuals. In this pa-per an alternative selection method, based on the technique of Random Forests, is proposed in the context of classification, with an application to a real dataset.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
AUSTIN, P. and TU, J. (2004): Bootstrap methods for developing predictive mod-els. The American Statistician, 58, 131–137.
BREIMAN, L., FRIEDMAN, J.H., OLSHEN, R.A. and STONE, C.J. (1984): Classification and Regression Trees. Chapman & Hall, London.
BREIMAN, L. (1996a): The heuristic of instability in model selection. Annals of Statistics, 24, 2350–2383.
BREIMAN, L. (1996b): Bagging predictions. Machine Learning, 24, 123–140.
BREIMAN, L. (2001a): Random Forests. Machine Learning, 45, 5–32.
BREIMAN, L. (2001b): Statistical modeling: the two cultures. Statistical Science, 16, 199–231.
BREIMAN, L. (2002): Manual on setting up, using, and understanding Random Forests v3.1. Technical Report, http://oz.berkeley.edu/users/breiman.
DIETTERICH, T. (2000): An experimental comparison of three methods for con-struction ensembles of decision trees: bagging, boosting and randomization. Machine Learning, 40, 139–157.
ENNIS, M., HINTON, G., NAYLOR, D., REVOW, M. and TIBSHIRANI, R. (1998): A comparison of statistical learning methods on the gusto database. Statistics in Medicine, 17, 2501–2508.
GUGLIELMI, A., RUZZENENTE, A., SANDRI, M., KIND, R., LOMBARDO, F., RODELLA, L., CATALANO, F., DE MANZONI, G. and CORDIANO, C. (2002): Risk assessment and prediction of rebleeding in bleeding gastroduodenal ulcer. Endoscopy, 34, 771–779.
HOCKING, R.R.. (1976): The analysis and selection of variables in linear regression. Biometrics, 42, 1–49.
MILLER, A.J. (1984): Selection of subsets of regression variables. Journal of the Royal Statistical Society, Series A, 147, 389–425.
SANDRI, M. and ZUCCOLOTTO, P. (2004): Classification with Random Forests: the theoretical framework. Rapporto di Ricerca del Dipartimento Metodi Quantitativi, Università degli Studi di Brescia, 235.
SANDRI, M. and ZUCCOLOTTO, P. (2006): Analysis of a bias effect on a tree-based variable importance measure. Evaluation of an empirical adjustment strategy. Manuscript.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Heidelberg
About this paper
Cite this paper
Sandri, M., Zuccolotto, P. (2006). Variable Selection Using Random Forests. In: Zani, S., Cerioli, A., Riani, M., Vichi, M. (eds) Data Analysis, Classification and the Forward Search. Studies in Classification, Data Analysis, and Knowledge Organization. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-35978-8_30
Download citation
DOI: https://doi.org/10.1007/3-540-35978-8_30
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-35977-7
Online ISBN: 978-3-540-35978-4
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)