scispace - formally typeset
Open AccessJournal ArticleDOI

Tuning parameter selection in high dimensional penalized likelihood

Reads0
Chats0
TLDR
In this article, the authors proposed to select the tuning parameter by optimizing the generalized information criterion with an appropriate model complexity penalty, which diverges at the rate of some power of ǫ(p) depending on the tail probability behavior of the response variables.
Abstract
Summary Determining how to select the tuning parameter appropriately is essential in penalized likelihood methods for high dimensional data analysis. We examine this problem in the setting of penalized likelihood methods for generalized linear models, where the dimensionality of covariates p is allowed to increase exponentially with the sample size n. We propose to select the tuning parameter by optimizing the generalized information criterion with an appropriate model complexity penalty. To ensure that we consistently identify the true model, a range for the model complexity penalty is identified in the generlized information criterion. We find that this model complexity penalty should diverge at the rate of some power of  log (p) depending on the tail probability behaviour of the response variables. This reveals that using the Akaike information criterion or Bayes information criterion to select the tuning parameter may not be adequate for consistently identifying the true model. On the basis of our theoretical study, we propose a uniform choice of the model complexity penalty and show that the approach proposed consistently identifies the true model among candidate models with asymptotic probability 1. We justify the performance of the procedure proposed by numerical simulations and a gene expression data analysis.

read more

Citations
More filters
Journal ArticleDOI

High-Dimensional Interaction Detection With False Sign Rate Control

TL;DR: It is proved that the examined method enjoys the same oracle inequalities as the lasso estimator and further admits an explicit bound on the false sign rate, which can be asymptotically vanishing in ultrahigh-dimensional regression models.
Journal ArticleDOI

A study on tuning parameter selection for the high-dimensional lasso

TL;DR: In this article, the authors develop a suite of information criteria for choosing the tuning parameter in lasso regression by leveraging the literature on high-dimensional variance estimation and derive intuition showing that existing information-theoretic approaches work poorly in this setting.
Posted Content

Analysis of Networks via the Sparse $\beta$-Model

TL;DR: This work proposes the Sparse $\beta$-Model, a new network model that interpolates the celebrated Erdős-Renyi model and the $\beta-model that assigns one different parameter to each node, and shows via a monotonicity lemma that the seemingly combinatorial computational problem due to the $\ell_0$-penalty can be overcome.
Journal ArticleDOI

A Fast and Scalable Implementation Method for Competing Risks Data with the R Package fastcmprsk

TL;DR: An R package is developed that uses a novel forward-backward scan algorithm to significantly reduce the computational complexity for parameter estimation by exploiting the structure of the subject-specific risk sets in the Fine-Gray model.
Journal ArticleDOI

Improved cancer biomarkers identification using network-constrained infinite latent feature selection.

TL;DR: A novel method that combines the infinite latent feature selection (ILFS) method with the functional interaction (FIs) network to rank the biomarkers and suggests that network-constrained ILFS can identify cancer-related genes with a higher discriminative power and biological significance.
References
More filters
Journal ArticleDOI

Regression Shrinkage and Selection via the Lasso

TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Journal ArticleDOI

Estimating the Dimension of a Model

TL;DR: In this paper, the problem of selecting one of a number of models of different dimensions is treated by finding its Bayes solution, and evaluating the leading terms of its asymptotic expansion.
Book

Generalized Linear Models

TL;DR: In this paper, a generalization of the analysis of variance is given for these models using log- likelihoods, illustrated by examples relating to four distributions; the Normal, Binomial (probit analysis, etc.), Poisson (contingency tables), and gamma (variance components).
Book

The Elements of Statistical Learning: Data Mining, Inference, and Prediction

TL;DR: In this paper, the authors describe the important ideas in these areas in a common conceptual framework, and the emphasis is on concepts rather than mathematics, with a liberal use of color graphics.
Book ChapterDOI

Information Theory and an Extension of the Maximum Likelihood Principle

TL;DR: In this paper, it is shown that the classical maximum likelihood principle can be considered to be a method of asymptotic realization of an optimum estimate with respect to a very general information theoretic criterion.