scispace - formally typeset
Open AccessJournal ArticleDOI

False Discoveries Occur Early on the Lasso Path

Weijie J. Su, +2 more
- 01 Oct 2017 - 
- Vol. 45, Iss: 5, pp 2133-2150
Reads0
Chats0
TLDR
It is demonstrated that true features and null features are always interspersed on the Lasso path, and that this phenomenon occurs no matter how strong the effect sizes are.
Abstract
In regression settings where explanatory variables have very low correlations and there are relatively few effects, each of large magnitude, we expect the Lasso to find the important variables with few errors, if any. This paper shows that in a regime of linear sparsity—meaning that the fraction of variables with a nonvanishing effect tends to a constant, however small—this cannot really be the case, even when the design variables are stochastically independent. We demonstrate that true features and null features are always interspersed on the Lasso path, and that this phenomenon occurs no matter how strong the effect sizes are. We derive a sharp asymptotic trade-off between false and true positive rates or, equivalently, between measures of type I and type II errors along the Lasso path. This trade-off states that if we ever want to achieve a type II error (false negative rate) under a critical value, then anywhere on the Lasso path the type I error (false positive rate) will need to exceed a given threshold so that we can never have both errors at a low level at the same time. Our analysis uses tools from approximate message passing (AMP) theory as well as novel elements to deal with a possibly adaptive selection of the Lasso regularizing parameter.

read more

Content maybe subject to copyright    Report

Citations
More filters
Posted ContentDOI

Accurate prediction of cell composition, age, smoking consumption and infection serostatus based on blood DNA methylation profiles

TL;DR: This study substantially improves predictions of blood cell composition based on methylation profiles, which will be critical in the emerging field of medical epigenomics, by using elastic net regularized and stability selected regression models to predict the circulating levels of 70 blood cell subsets.
MonographDOI

A Unifying Tutorial on Approximate Message Passing

TL;DR: Approximate Message Passing (AMP) algorithms have become extremely popular in various structured high-dimensional statistical problems and have been extended for use in computer science and machine learning as mentioned in this paper .
Posted Content

SLOPE for Sparse Linear Regression:Asymptotics and Optimal Regularization

TL;DR: In this article, the SLOPE estimator generalizes LASSO by penalizing different coordinates of the estimate according to their magnitudes, and a computational feasible way to optimally design the regularizing sequences such that the fundamental limits are reached.
Journal ArticleDOI

Nested model averaging on solution path for high-dimensional linear regression

TL;DR: This work proposes to combine model averaging with regularized estimators (e.g., lasso, elastic net, and Sorted L‐One Penalized Estimation [SLOPE]) on the solution path for high‐dimensional linear regression.
Proceedings Article

The Complete Lasso Tradeoff Diagram

TL;DR: In this paper, the tradeoff between false discovery rate (FDR) and power in variable selection was studied in a regime of linear sparsity under random designs, and a complete Lasso tradeoff diagram was proposed.
References
More filters
Journal ArticleDOI

Regression Shrinkage and Selection via the Lasso

TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Journal ArticleDOI

Regularization and variable selection via the elastic net

TL;DR: It is shown that the elastic net often outperforms the lasso, while enjoying a similar sparsity of representation, and an algorithm called LARS‐EN is proposed for computing elastic net regularization paths efficiently, much like algorithm LARS does for the lamba.
Journal ArticleDOI

Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties

TL;DR: In this article, penalized likelihood approaches are proposed to handle variable selection problems, and it is shown that the newly proposed estimators perform as well as the oracle procedure in variable selection; namely, they work as well if the correct submodel were known.
Journal ArticleDOI

Model selection and estimation in regression with grouped variables

TL;DR: In this paper, instead of selecting factors by stepwise backward elimination, the authors focus on the accuracy of estimation and consider extensions of the lasso, the LARS algorithm and the non-negative garrotte for factor selection.
Related Papers (5)