scispace - formally typeset
Open AccessJournal ArticleDOI

"Pre-conditioning" for feature selection and regression in high-dimensional problems

TLDR
In this article, the authors proposed a method for variable selection that first estimates the regression function, yielding a "pre-conditioned" response variable, and then applies a standard procedure such as forward stepwise selection or the LASSO to the preconditioned response variable.
Abstract
We consider regression problems where the number of predictors greatly exceeds the number of observations. We propose a method for variable selection that first estimates the regression function, yielding a "pre-conditioned" response variable. The primary method used for this initial regression is supervised principal components. Then we apply a standard procedure such as forward stepwise selection or the LASSO to the pre-conditioned response variable. In a number of simulated and real data examples, this two-step procedure outperforms forward stepwise selection or the usual LASSO (applied directly to the raw outcome). We also show that under a certain Gaussian latent variable model, application of the LASSO to the pre-conditioned response variable is consistent as the number of predictors and observations increases. Moreover, when the observational noise is rather large, the suggested procedure can give a more accurate estimate than LASSO. We illustrate our method on some real problems, including survival analysis with microarray data.

read more

Citations
More filters
Journal ArticleDOI

Sure independence screening for ultrahigh dimensional feature space

TL;DR: In this article, the authors introduce the concept of sure screening and propose a sure screening method that is based on correlation learning, called sure independence screening, to reduce dimensionality from high to a moderate scale that is below the sample size.
Posted Content

Sure Independence Screening for Ultra-High Dimensional Feature Space

TL;DR: The concept of sure screening is introduced and a sure screening method that is based on correlation learning, called sure independence screening, is proposed to reduce dimensionality from high to a moderate scale that is below the sample size.
Journal Article

Ultrahigh Dimensional Feature Selection: Beyond The Linear Model

TL;DR: This paper extends ISIS, without explicit definition of residuals, to a general pseudo-likelihood framework, which includes generalized linear models as a special case and improves ISIS by allowing feature deletion in the iterative process.
Journal ArticleDOI

Robust rank correlation based screening

TL;DR: In this article, a robust rank correlation screening (RRCS) method is proposed to deal with ultra-high dimensional data, which is based on the Kendall \tau correlation coefficient between response and predictor variables rather than the Pearson correlation.
Journal ArticleDOI

Recent advances and emerging challenges of feature selection in the context of big data

TL;DR: The origins and importance of feature selection are discussed and recent contributions in a range of applications are outlined, from DNA microarray analysis to face recognition.
References
More filters
Journal ArticleDOI

Regression Shrinkage and Selection via the Lasso

TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Journal ArticleDOI

Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties

TL;DR: In this article, penalized likelihood approaches are proposed to handle variable selection problems, and it is shown that the newly proposed estimators perform as well as the oracle procedure in variable selection; namely, they work as well if the correct submodel were known.
Journal ArticleDOI

The adaptive lasso and its oracle properties

TL;DR: A new version of the lasso is proposed, called the adaptive lasso, where adaptive weights are used for penalizing different coefficients in the ℓ1 penalty, and the nonnegative garotte is shown to be consistent for variable selection.
Related Papers (5)