scispace - formally typeset
Search or ask a question

Showing papers on "Latent variable model published in 2010"


Journal ArticleDOI
TL;DR: Partial least squares (PLS) regression as mentioned in this paper is a recent technique that combines features from and generalizes principal component analysis (PCA) and multiple linear regression, which can be used to predict a set of dependent variables from a subset of independent variables or predictors.
Abstract: Partial least squares (PLS) regression (a.k.a. projection on latent structures) is a recent technique that combines features from and generalizes principal component analysis (PCA) and multiple linear regression. Its goal is to predict a set of dependent variables from a set of independent variables or predictors. This prediction is achieved by extracting from the predictors a set of orthogonal factors called latent variables which have the best predictive power. These latent variables can be used to create displays akin to PCA displays. The quality of the prediction obtained from a PLS regression model is evaluated with cross-validation techniques such as the bootstrap and jackknife. There are two main variants of PLS regression: The most common one separates the roles of dependent and independent variables; the second one—used mostly to analyze brain imaging data—gives the same roles to dependent and independent variables. Copyright © 2010 John Wiley & Sons, Inc. For further resources related to this article, please visit the WIREs website.

1,062 citations


Journal ArticleDOI
TL;DR: A method to visualize comorbidity networks is proposed and it is argued that this approach generates realistic hypotheses about pathways to comor bidity, overlapping symptoms, and diagnostic boundaries, that are not naturally accommodated by latent variable models.
Abstract: The pivotal problem of comorbidity research lies in the psychometric foundation it rests on, that is, latent variable theory, in which a mental disorder is viewed as a latent variable that causes a constellation of symptoms. From this perspective, comorbidity is a (bi)directional relationship between multiple latent variables. We argue that such a latent variable perspective encounters serious problems in the study of comorbidity, and offer a radically different conceptualization in terms of a network approach, where comorbidity is hypothesized to arise from direct relations between symptoms of multiple disorders. We propose a method to visualize comorbidity networks and, based on an empirical network for major depression and generalized anxiety, we argue that this approach generates realistic hypotheses about pathways to comorbidity, overlapping symptoms, and diagnostic boundaries, that are not naturally accommodated by latent variable models: Some pathways to comorbidity through the symptom space are more likely than others; those pathways generally have the same direction (i.e., from symptoms of one disorder to symptoms of the other); overlapping symptoms play an important role in comorbidity; and boundaries between diagnostic categories are necessarily fuzzy.

918 citations


Proceedings ArticleDOI
26 Apr 2010
TL;DR: This work develops an unsupervised model to estimate relationship strength from interaction activity and user similarity and evaluates it on real-world data from Facebook and LinkedIn, showing that the estimated link weights result in higher autocorrelation and lead to improved classification accuracy.
Abstract: Previous work analyzing social networks has mainly focused on binary friendship relations. However, in online social networks the low cost of link formation can lead to networks with heterogeneous relationship strengths (e.g., acquaintances and best friends mixed together). In this case, the binary friendship indicator provides only a coarse representation of relationship information. In this work, we develop an unsupervised model to estimate relationship strength from interaction activity (e.g., communication, tagging) and user similarity. More specifically, we formulate a link-based latent variable model, along with a coordinate ascent optimization procedure for the inference. We evaluate our approach on real-world data from Facebook and LinkedIn, showing that the estimated link weights result in higher autocorrelation and lead to improved classification accuracy.

725 citations


Proceedings Article
09 Oct 2010
TL;DR: A multi-level generative model that reasons jointly about latent topics and geographical regions is presented, which recovers coherent topics and their regional variants, while identifying geographic areas of linguistic consistency.
Abstract: The rapid growth of geotagged social media raises new computational possibilities for investigating geographic linguistic variation. In this paper, we present a multi-level generative model that reasons jointly about latent topics and geographical regions. High-level topics such as "sports" or "entertainment" are rendered differently in each geographic region, revealing topic-specific regional distinctions. Applied to a new dataset of geotagged microblogs, our model recovers coherent topics and their regional variants, while identifying geographic areas of linguistic consistency. The model also enables prediction of an author's geographic location from raw text, outperforming both text regression and supervised topic models.

691 citations


Book ChapterDOI
01 Jan 2010
TL;DR: A statistical model can be called a latent class (LC) or mixture model if it assumes that some of its parameters differ across unobserved subgroups, LCs, or mixture components as mentioned in this paper.
Abstract: A statistical model can be called a latent class (LC) or mixture model if it assumes that some of its parameters differ across unobserved subgroups, LCs, or mixture components. This rather general idea has several seemingly unrelated applications, the most important of which are clustering, scaling, density estimation, and random-effects modeling. This article describes simple LC models for clustering, restricted LC models for scaling, and mixture regression models for nonparametric random-effects modeling, as well as gives an overview of recent developments in the field of LC analysis. Moreover, attention is paid to topics such as maximum likelihood estimation, identification issues, model selection, and software.

431 citations


Proceedings Article
31 Mar 2010
TL;DR: In this article, a variational inference framework for training the Gaussian process latent variable model and thus performing Bayesian nonlinear dimensionality reduction is introduced, which can automatically select the dimensionality of the nonlinear latent space.
Abstract: We introduce a variational inference framework for training the Gaussian process latent variable model and thus performing Bayesian nonlinear dimensionality reduction. This method allows us to variationally integrate out the input variables of the Gaussian process and compute a lower bound on the exact marginal likelihood of the nonlinear latent variable model. The maximization of the variational lower bound provides a Bayesian training procedure that is robust to overfitting and can automatically select the dimensionality of the nonlinear latent space. We demonstrate our method on real world datasets. The focus in this paper is on dimensionality reduction problems, but the methodology is more general. For example, our algorithm is immediately applicable for training Gaussian process models in the presence of missing or uncertain inputs.

338 citations


Journal ArticleDOI
TL;DR: A two-tier item factor analysis model is developed that reduces the dimensionality of the latent variable space, and consequently significant computational savings, and an EM algorithm for full-information maximum marginal likelihood estimation is developed.
Abstract: Motivated by Gibbons et al.’s (Appl. Psychol. Meas. 31:4–19, 2007) full-information maximum marginal likelihood item bifactor analysis for polytomous data, and Rijmen, Vansteelandt, and De Boeck’s (Psychometrika 73:167–182, 2008) work on constructing computationally efficient estimation algorithms for latent variable models, a two-tier item factor analysis model is developed in this research. The modeling framework subsumes standard multidimensional IRT models, bifactor IRT models, and testlet response theory models as special cases. Features of the model lead to a reduction in the dimensionality of the latent variable space, and consequently significant computational savings. An EM algorithm for full-information maximum marginal likelihood estimation is developed. Simulations and real data demonstrations confirm the accuracy and efficiency of the proposed methods. Three real data sets from a large-scale educational assessment, a longitudinal public health survey, and a scale development study measuring patient reported quality of life outcomes are analyzed as illustrations of the model’s broad range of applicability.

226 citations


Journal ArticleDOI
TL;DR: The goal of this paper is to discover the objects present in the images by analyzing unlabeled data and searching for re-occurring patterns, and a rigorous framework for evaluating unsupervised object discovery methods is proposed.
Abstract: The goal of this paper is to evaluate and compare models and methods for learning to recognize basic entities in images in an unsupervised setting. In other words, we want to discover the objects present in the images by analyzing unlabeled data and searching for re-occurring patterns. We experiment with various baseline methods, methods based on latent variable models, as well as spectral clustering methods. The results are presented and compared both on subsets of Caltech256 and MSRC2, data sets that are larger and more challenging and that include more object classes than what has previously been reported in the literature. A rigorous framework for evaluating unsupervised object discovery methods is proposed.

216 citations


Proceedings ArticleDOI
13 Jun 2010
TL;DR: A new class of probabilistic latent variable model called the Implicit Mixture of Conditional Restricted Boltzmann Machines (imCRBM) for use in human pose tracking and its use in the context of Bayesian filtering for multi-view and monocular pose tracking is demonstrated.
Abstract: We introduce a new class of probabilistic latent variable model called the Implicit Mixture of Conditional Restricted Boltzmann Machines (imCRBM) for use in human pose tracking. Key properties of the imCRBM are as follows: (1) learning is linear in the number of training exemplars so it can be learned from large datasets; (2) it learns coherent models of multiple activities; (3) it automatically discovers atomic “movemes” and (4) it can infer transitions between activities, even when such transitions are not present in the training set. We describe the model and how it is learned and we demonstrate its use in the context of Bayesian filtering for multi-view and monocular pose tracking. The model handles difficult scenarios including multiple activities and transitions among activities. We report state-of-the-art results on the HumanEva dataset.

164 citations


Proceedings ArticleDOI
06 Aug 2010
TL;DR: The modeling framework can be viewed as a combination of dimensionality reduction and graphical modeling (to capture remaining statistical structure not attributable to the latent variables) and it consistently estimates both the number of hidden components and the conditional graphical model structure among the observed variables.
Abstract: Suppose we have samples of a subset of a collection of random variables. No additional information is provided about the number of latent variables, nor of the relationship between the latent and observed variables. Is it possible to discover the number of hidden components, and to learn a statistical model over the entire collection of variables? We address this question in the setting in which the latent and observed variables are jointly Gaussian, with the conditional statistics of the observed variables conditioned on the latent variables being specified by a graphical model. As a first step we give natural conditions under which such latent-variable Gaussian graphical models are identifiable given marginal statistics of only the observed variables. Essentially these conditions require that the conditional graphical model among the observed variables is sparse, while the effect of the latent variables is “spread out” over most of the observed variables. Next we propose a tractable convex program based on regularized maximum-likelihood for model selection in this latent-variable setting; the regularizer uses both the l 1 norm and the nuclear norm. Our modeling framework can be viewed as a combination of dimensionality reduction (to identify latent variables) and graphical modeling (to capture remaining statistical structure not attributable to the latent variables), and it consistently estimates both the number of hidden components and the conditional graphical model structure among the observed variables. These results are applicable in the high-dimensional setting in which the number of latent/observed variables grows with the number of samples of the observed variables. The geometric properties of the algebraic varieties of sparse matrices and of low-rank matrices play an important role in our analysis.

157 citations


Journal ArticleDOI
TL;DR: In this article, two methods have been proposed: the sequential approach, in which the latent variables are built before their integration with the traditional explanatory variables in the choice model and the simultaneous approach, which both processes are done together, albeit with a sophisticated but fairly complex treatment.
Abstract: The formulation of hybrid discrete choice models, including both observable alternative attributes and latent variables associated with attitudes and perceptions, has become a topic of discussion once more To estimate models integrating both kinds of variables, two methods have been proposed: the sequential approach, in which the latent variables are built before their integration with the traditional explanatory variables in the choice model and the simultaneous approach, in which both processes are done together, albeit with a sophisticated but fairly complex treatment Here both approaches are applied to estimate hybrid choice models by using two data sets: one from the Santiago Panel (an urban mode choice context with many alternatives) and another consisting of synthetic data Differences between both approaches were found as well as similarities not found in earlier studies Even when both approaches result in unbiased estimators, problems arise when valuations are obtained such as the value of tim

Book Chapter
01 Jan 2010
TL;DR: A Bayesian nonparametric model uses only a finite subset of the available parameter dimensions to explain a finite sample of observations, such that the effective complexity of the model adapts to the data.
Abstract: A Bayesian nonparametric model is a Bayesian model on an infinite-dimensional parameter space. The parameter space is typically chosen as the set of all possible solutions for a given learning problem. For example, in a regression problem the parameter space can be the set of continuous functions, and in a density estimation problem the space can consist of all densities. A Bayesian nonparametric model uses only a finite subset of the available parameter dimensions to explain a finite sample of observations, with the set of dimensions chosen depending on the sample, such that the effective complexity of the model (as measured by the number of dimensions used) adapts to the data. Classical adaptive problems, such as nonparametric estimation and model selection, can thus be formulated as Bayesian inference problems. Popular examples of Bayesian nonparametric models include Gaussian process regression, in which the correlation structure is refined with growing sample size, and Dirichlet process mixture models for clustering, which adapt the number of clusters to the complexity of the data. Bayesian nonparametric models have recently been applied to a variety of machine learning problems, including regression, classification, clustering, latent variable modeling, sequential modeling, image segmentation, source separation and grammar induction.

Journal ArticleDOI
TL;DR: An approach to consistency testing is presented that builds on prior work demonstrating that parallel analyses of categorical and dimensional comparison data provide an accurate index of the relative fit of competing structural models.
Abstract: A number of recent studies have used Meehl's (1995) taxometric method to determine empirically whether one should model assessment-related constructs as categories or dimensions. The taxometric method includes multiple data-analytic procedures designed to check the consistency of results. The goal is to differentiate between strong evidence of categorical structure, strong evidence of dimensional structure, and ambiguous evidence that suggests withholding judgment. Many taxometric consistency tests have been proposed, but their use has not been operationalized and studied rigorously. What tests should be performed, how should results be combined, and what thresholds should be applied? We present an approach to consistency testing that builds on prior work demonstrating that parallel analyses of categorical and dimensional comparison data provide an accurate index of the relative fit of competing structural models. Using a large simulation study spanning a wide range of data conditions, we examine many critical elements of this approach. The results provide empirical support for what marks the first rigorous operationalization of consistency testing. We discuss and empirically illustrate guidelines for implementing this approach and suggest avenues for future research to extend the practice of consistency testing to other techniques for modeling latent variables in the realm of psychological assessment.

Journal ArticleDOI
TL;DR: In this article, a method for interval estimation of scale reliability with discrete data is outlined, which is applicable with multi-item instruments consisting of binary measures, and is developed within the latent variable modeling methodology.
Abstract: A method for interval estimation of scale reliability with discrete data is outlined. The approach is applicable with multi-item instruments consisting of binary measures, and is developed within the latent variable modeling methodology. The procedure is useful for evaluation of consistency of single measures and of sum scores from item sets following the 2-parameter logistic model or the 1-parameter logistic model. An extension of the method is described for constructing confidence intervals of change in reliability due to instrument revision. The proposed procedure is illustrated with an example.

Book ChapterDOI
01 Jan 2010
TL;DR: The multigroup latent class (LC) model as mentioned in this paper is an extension of the standard LC model for the analysis of latent structures of observed categorical variables across two or more groups.
Abstract: This chapter introduces the basic multigroup latent class (LC) model, and discusses two important extensions of the basic model, an extension for dealing with ordinal indicators and for modeling the latent variables as ordinal variables. It analyses measurement invariance using multigroup LC models, discussing the general procedure as well as methods for parameter estimation and evaluation of model fit. The multigroup extension of the standard LC model has been developed for the analysis of latent structures of observed categorical variables across two or more groups. The chapter presents three parameterizations of the multigroup LC models: probabilistic, log-linear, and logistic parameterizations. Multigroup LC models assume the presence of three types of categorical variables: observed variables; an unobserved variable that accounts for the relationships between the observed variables. LC models are usually estimated by means of maximum-likelihood under the assumption of a multinomial distribution for the indicator variables in the model.

01 Jan 2010
TL;DR: This paper proposed a generative dependency parsing model which uses binary latent variables to induce conditioning features, which achieved state-of-the-art results on three different languages, and showed that the features induced by the ISBN's latent variables are crucial to this success.
Abstract: We propose a generative dependency parsing model which uses binary latent variables to induce conditioning features. To define this model we use a recently proposed class of Bayesian Networks for structured prediction, Incremental Sigmoid Belief Networks. We demonstrate that the proposed model achieves state-of-the-art results on three different languages. We also demonstrate that the features induced by the ISBN's latent variables are crucial to this success, and show that the proposed model is particularly good on long dependencies.

Journal ArticleDOI
TL;DR: H Hancock and Karen M Samuelson as discussed by the authors published a volume of the Information Age 2008, 370 pages, $3999 (paperback) which is the result of a conference proceedings.
Abstract: Gregory R Hancock & Karen M Samuelson (Eds) Charlotte, NC: Information Age, 2008, 370 pages, $3999 (paperback) This volume, although not precisely a conference proceedings, is the result of a

Journal ArticleDOI
TL;DR: Similar prediction patterns were found for physical activity as well as fruit and vegetable intake: changes in intention and self-efficacy predicted changes in planning, which in turn corresponded to changes in behavior.
Abstract: Can latent true changes in intention, planning, and self-efficacy account for latent change in two health behaviors (physical activity as well as fruit and vegetable intake)? Baseline data on predictors and behaviors and corresponding follow-up data four weeks later were collected from 853 participants. Interindividual differences in change and change-change associations were analyzed using structural equation modeling. For both behaviors, similar prediction patterns were found: changes in intention and self-efficacy predicted changes in planning, which in turn corresponded to changes in behavior. This evidence confirms that change predicts change, which is an inherent precondition in behavior change theories.

Journal ArticleDOI
TL;DR: This paper provides an introduction to a recently developed conceptual framework-the dimensional-categorical spectrum-for utilizing general factor mixture models to explore the latent structures of psychological constructs.
Abstract: This paper provides an introduction to a recently developed conceptual framework— the dimensional–categorical spectrum—for utilizing general factor mixture models to explore the latent structures of psychological constructs. This framework offers advantages over traditional latent variable models that usually employ either continuous latent factors or categorical latent class variables to characterize the latent structure and require an a priori assumption about the underlying nature of the construct as either purely dimension or purely categorical. The modeling process is discussed in detail and then illustrated with data on the delinquency items of Achenbach’s child behavior checklist from a sample of children in the National Adolescent and Child Treatment Study.

Journal ArticleDOI
TL;DR: In this paper, a mixture item response theory model has been suggested as a potentially useful methodology for identifying latent groups formed along secondary, possibly nuisance dimensions, which can be used to identify latent groups.
Abstract: Mixture item response theory models have been suggested as a potentially useful methodology for identifying latent groups formed along secondary, possibly nuisance dimensions. In this article, we d...

Book ChapterDOI
05 Sep 2010
TL;DR: A novel probabilistic inference algorithm for 3D shape estimation is proposed by maximum likelihood estimates of the GPLVM latent variables and the camera parameters that best fit generated 3D shapes to given silhouettes.
Abstract: In this paper we propose a probabilistic framework that models shape variations and infers dense and detailed 3D shapes from a single silhouette. We model two types of shape variations, the object phenotype variation and its pose variation using two independent Gaussian Process Latent Variable Models (GPLVMs) respectively. The proposed shape variation models are learnt from 3D samples without prior knowledge about object class, e.g. object parts and skeletons, and are combined to fully span the 3D shape space. A novel probabilistic inference algorithm for 3D shape estimation is proposed by maximum likelihood estimates of the GPLVM latent variables and the camera parameters that best fit generated 3D shapes to given silhouettes. The proposed inference involves a small number of latent variables and it is computationally efficient. Experiments on both human body and shark data demonstrate the efficacy of our new approach.

Book ChapterDOI
01 Dec 2010
TL;DR: In this paper, a set of applications of one class of longitudinal growth analysis - latent curve (LCM) and latent change score (LCS) analysis using structural equation modeling (SEM) techniques are described.
Abstract: This paper describes a set of applications of one class of longitudinal growth analysis - latent curve (LCM) and latent change score (LCS) analysis using structural equation modeling (SEM) techniques. These techniques are organized in five sections based on Baltes & Nesselroade (1979). (1) Describing the observed and unobserved longitudinal data. (2) Characterizing the developmental shape of both individuals and groups. (3) Examining the predictors of individual and group differences in developmental shapes. (4) Studying dynamic determinants among variables over time. (5) Studying group differences in dynamic determinants among variables over time. To illustrate all steps, we present SEM analyses of a relatively large set of data from the National Longitudinal Survey of Youth (NLSY). The inclusion of all five aspects of latent curve modeling is not often used in longitudinal analyses, so we discuss why more efforts to include all five are needed in developmental research.

Journal ArticleDOI
TL;DR: In this paper, two latent variable model predictive control (LV-MPC) algorithms are developed for trajectory tracking and disturbance rejection in batch processes, based on multi-phase PCA models developed using batch-wise unfolding of batch data arrays.

Journal ArticleDOI
TL;DR: A broad class of semiparametric Bayesian SEMs, which allow mixed categorical and continuous manifest variables while also allowing the latent variables to have unknown distributions is proposed, based on centered Dirichlet process and CDP mixture models.
Abstract: Structural equation models (SEMs) with latent variables are widely useful for sparse covariance structure modeling and for inferring relationships among latent variables. Bayesian SEMs are appealing in allowing for the incorporation of prior information and in providing exact posterior distributions of unknowns, including the latent variables. In this article, we propose a broad class of semiparametric Bayesian SEMs, which allow mixed categorical and continuous manifest variables while also allowing the latent variables to have unknown distributions. In order to include typical identifiability restrictions on the latent variable distributions, we rely on centered Dirichlet process (CDP) and CDP mixture (CDPM) models. The CDP will induce a latent class model with an unknown number of classes, while the CDPM will induce a latent trait model with unknown densities for the latent traits. A simple and efficient Markov chain Monte Carlo algorithm is developed for posterior computation, and the methods are illustrated using simulated examples, and several applications.

Proceedings ArticleDOI
25 Jul 2010
TL;DR: A novel topic model using the Pitman-Yor(PY) process is proposed, called the PY topic model, which captures two properties of a document; a power-law word distribution and the presence of multiple topics.
Abstract: One important approach for knowledge discovery and data mining is to estimate unobserved variables because latent variables can indicate hidden specific properties of observed data. The latent factor model assumes that each item in a record has a latent factor; the co-occurrence of items can then be modeled by latent factors. In document modeling, a record indicates a document represented as a "bag of words," meaning that the order of words is ignored, an item indicates a word and a latent factor indicates a topic. Latent Dirichlet allocation (LDA) is a widely used Bayesian topic model applying the Dirichlet distribution over the latent topic distribution of a document having multiple topics. LDA assumes that latent topics, i.e., discrete latent variables, are distributed according to a multinomial distribution whose parameters are generated from the Dirichlet distribution. LDA also models a word distribution by using a multinomial distribution whose parameters follows the Dirichlet distribution. This Dirichlet-multinomial setting, however, cannot capture the power-law phenomenon of a word distribution, which is known as Zipf's law in linguistics. We therefore propose a novel topic model using the Pitman-Yor(PY) process, called the PY topic model. The PY topic model captures two properties of a document; a power-law word distribution and the presence of multiple topics. In an experiment using real data, this model outperformed LDA in document modeling in terms of perplexity.

Journal ArticleDOI
TL;DR: A structure theory with emphasis on the zeroless case is presented, which is generic in the setting considered and the latent variables are modeled as a possibly singular autoregressive process and (generalized) Yule Walker equations are used for parameter estimation.

Journal ArticleDOI
TL;DR: In this paper, the authors show that the appropriate parameter estimates are easily formulated from parameter estimates commonly available from existing structural equation modeling (SEM) software packages, including the main and interaction effects are scale-free, as are the factor loadings.
Abstract: Standardized parameter estimates are routinely used to summarize the results of multiple regression models of manifest variables and structural equation models of latent variables, because they facilitate interpretation. Although the typical standardization of interaction terms is not appropriate for multiple regression models, straightforward alternatives are well known (Aiken & West, 1991; Friedrich, 1982). Whereas the analogous problem exists for the estimation of latent interactions in structural equation modeling (SEM), the problem is more complex and apparently has not been resolved. Here we demonstrate that the appropriate “standardized” parameter estimates are easily formulated from parameter estimates routinely available from existing SEM software packages. Some properties of the appropriate “standardized” solution are mathematically derived, including the demonstration that the main and interaction effects are scale-free, as are the factor loadings. These desirable properties of the standardized...

Journal ArticleDOI
TL;DR: Methods based on discrete random effects distributions and mixture models for application in recurrent measures and multivariate outcomes and sub-groups of patients who vary in treatment responsiveness are reviewed.
Abstract: Repeated measures and multivariate outcomes are an increasingly common feature of trials. Their joint analysis by means of random effects and latent variable models is appealing but patterns of heterogeneity in outcome profile may not conform to standard multivariate normal assumptions. In addition, there is much interest in both allowing for and identifying sub-groups of patients who vary in treatment responsiveness. We review methods based on discrete random effects distributions and mixture models for application in this field.


Journal ArticleDOI
TL;DR: A Bayesian model for a point-referenced spatially correlated ordered categorical response and methodology for inference is proposed and demonstrated to offer superior predictive performance as compared to a non-spatial cumulative probit model and a more standard Bayesian generalized linear spatial model.