In this paper we consider the problem of building a linear prediction model when the number of candidate predictors is large and the data possibly contains anomalies that are difficult to visualize and clean. We aim at predicting the non-outlying cases. Therefore, we need a method that is robust and scalable at the same time. We consider the stepwise algorithm LARS which is computationally very efficient but sensitive to outliers. We introduce two different approaches to robustify LARS. The plug-in approach replaces the classical correlations in LARS by robust correlation estimates. The cleaning approach first transforms the dataset by shrinking the outliers toward the bulk of the data (which we call multivariate Winsorization) and then app...
Large datasets are more and more common in many research fields. In particular, in the linear regres...
Problem statement: Bootstrap approach had introduced new advancement in modeling and model evaluatio...
The problem of selecting a parsimonious subset of variables from a large number of predictors in a r...
In this paper we consider the problem of building a linear prediction model when the number of candi...
This study considers the problem of building a linear prediction model when the number of candidate ...
In variable selection problems, when the number of candidate covariates is relatively large, the "tw...
The least angle regression selection (LARS) algorithms that use the classical sample means, variance...
We consider the problem of selecting a parsimonious subset of explanatory variables from a potential...
Least Angle Regression(LARS)is a variable selection method with proven performance for cross-section...
The purpose of model selection algorithms such as All Subsets, Forward Selection, and Backward Elimi...
Classical step-by-step algorithms, such as forward selectio n (FS) and stepwise (SW) methods, are co...
Large datasets are more and more common in many research fields. In particular, in the linear regres...
Many regression problems exhibit a natural grouping among predictor variables. Examples are groups o...
© 2017 The Author(s) Published by the Royal Society. All rights reserved. Least angle regression, as...
Large datasets are more and more common in many research fields. In particular, in the linear regres...
Problem statement: Bootstrap approach had introduced new advancement in modeling and model evaluatio...
The problem of selecting a parsimonious subset of variables from a large number of predictors in a r...
In this paper we consider the problem of building a linear prediction model when the number of candi...
This study considers the problem of building a linear prediction model when the number of candidate ...
In variable selection problems, when the number of candidate covariates is relatively large, the "tw...
The least angle regression selection (LARS) algorithms that use the classical sample means, variance...
We consider the problem of selecting a parsimonious subset of explanatory variables from a potential...
Least Angle Regression(LARS)is a variable selection method with proven performance for cross-section...
The purpose of model selection algorithms such as All Subsets, Forward Selection, and Backward Elimi...
Classical step-by-step algorithms, such as forward selectio n (FS) and stepwise (SW) methods, are co...
Large datasets are more and more common in many research fields. In particular, in the linear regres...
Many regression problems exhibit a natural grouping among predictor variables. Examples are groups o...
© 2017 The Author(s) Published by the Royal Society. All rights reserved. Least angle regression, as...
Large datasets are more and more common in many research fields. In particular, in the linear regres...
Problem statement: Bootstrap approach had introduced new advancement in modeling and model evaluatio...
The problem of selecting a parsimonious subset of variables from a large number of predictors in a r...