scispace - formally typeset
Journal ArticleDOI

Selection of ordinally scaled independent variables with applications to international classification of functioning core sets

Reads0
Chats0
TLDR
In this article, the ordinal structure is taken into account by use of a difference penalty on adjacent dummy coefficients, and an alternative blockwise boosting procedure is proposed to select ordinally scaled independent variables in the classical linear model.
Abstract
Summary. Ordinal categorial variables arise commonly in regression modelling. Although the analysis of ordinal response variables has been well investigated, less work has been done concerning ordinal predictors. We consider so-called international classfication of functioning core sets for chronic widespread pain, in which many ordinal covariates are collected. The effect of specific international classification of functioning variables on a subjective measure of physical health is investigated, which requires strategies for variable selection. In this context, we propose methods for the selection of ordinally scaled independent variables in the classical linear model. The ordinal structure is taken into account by use of a difference penalty on adjacent dummy coefficients. It is shown how the group lasso can be used for the selection of ordinal predictors, and an alternative blockwise boosting procedure is proposed. Both methods are discussed in general, and applied to international classification of functioning core sets for chronic widespread pain.

read more

Citations
More filters
Book

Regression for Categorical Data

TL;DR: This paper presents a meta-modelling architecture for binary regression that combines log-linear and graphical models, and some basic tools for random effects modeling are presented.
Journal ArticleDOI

Towards a minimal generic set of domains of functioning and health.

TL;DR: The minimal generic set proposed in this study represents the first step in developing a common metric of health to link information from the general population to information about sub-populations, such as clinical and institutionalized populations.
Journal ArticleDOI

Approaches to Regularized Regression – A Comparison between Gradient Boosting and the Lasso

TL;DR: Although following different strategies with respect to optimization and regularization, both methods imply similar constraints to the estimation problem leading to a comparable performance regarding prediction accuracy and variable selection in practice.
Journal ArticleDOI

Modelling monotonic effects of ordinal predictors in Bayesian regression models.

TL;DR: A Bayesian estimation method for monotonic effects is developed which allows us to incorporate prior information and to check the assumption of monotonicity, and is implemented in the R package brms, so that fitting monotony effects in a fully Bayesian framework is now straightforward.
References
More filters
Journal ArticleDOI

Random Forests

TL;DR: Internal estimates monitor error, strength, and correlation and these are used to show the response to increasing the number of features used in the forest, and are also applicable to regression.
Journal ArticleDOI

Regression Shrinkage and Selection via the Lasso

TL;DR: A new method for estimation in linear models called the lasso, which minimizes the residual sum of squares subject to the sum of the absolute value of the coefficients being less than a constant, is proposed.
Journal ArticleDOI

The MOS 36-item short-form health survey (SF-36). I. Conceptual framework and item selection.

John E. Ware, +1 more
- 01 Jun 1992 - 
TL;DR: A 36-item short-form survey designed for use in clinical practice and research, health policy evaluations, and general population surveys to survey health status in the Medical Outcomes Study is constructed.
Journal ArticleDOI

Regularization and variable selection via the elastic net

TL;DR: It is shown that the elastic net often outperforms the lasso, while enjoying a similar sparsity of representation, and an algorithm called LARS‐EN is proposed for computing elastic net regularization paths efficiently, much like algorithm LARS does for the lamba.