The Strength of Weak Learnability
Machine Learning
Prediction games and arcing algorithms
Neural Computation
Computational Statistics & Data Analysis
Computational Statistics & Data Analysis
Practical variable selection for generalized additive models
Computational Statistics & Data Analysis
Consistency of support vector machines using additive kernels for additive models
Computational Statistics & Data Analysis
Multinomial logit models with implicit variable selection
Advances in Data Analysis and Classification
Hi-index | 0.03 |
There are several possible approaches to combining ridge regression with boosting techniques. In the simple or naive approach the ridge estimator is used to fit iteratively the current residuals yielding an alternative to the usual ridge estimator. In partial boosting only part of the regression parameters are reestimated within one step of the iterative procedure. The technique allows to distinguish between mandatory variables that are always included in the analysis and optional variables that are chosen only if relevant. The resulting procedure selects optional variables in a similar way as the Lasso, yielding a reduced set of influential variables, while allowing for regularized estimation of the mandatory parameters. The suggested procedures are investigated within the classical framework of continuous response variables as well as in the case of generalized linear models. The performance in terms of prediction and the identification of relevant variables is compared to several competitors as the Lasso and the more recently proposed elastic net. For the evaluation of the identification of relevant variables pseudo ROC curves are introduced.