scispace - formally typeset
Search or ask a question

Showing papers in "Psychometrika in 1991"


Journal ArticleDOI
TL;DR: The use of kernel smoothing is explored, which is particularly well suited to this application of option characteristic curve estimation and lends itself to several interesting extensions.
Abstract: The option characteristic curve, the relation between ability and probability of choosing a particular option for a test item, can be estimated by nonparametric smoothing techniques. What is smoothed is the relation between some function of estimated examinee ability rankings and the binary variable indicating whether or not the option was chosen. This paper explores the use of kernel smoothing, which is particularly well suited to this application. Examples show that, with some help from the fast Fourier transform, estimates can be computed about 500 times as rapidly as when using commonly used parametric approaches such as maximum marginal likelihood estimation using the three-parameter logistic distribution. Simulations suggest that there is no loss of efficiency even when the population curves are three-parameter logistic. The approach lends itself to several interesting extensions.

330 citations


Journal ArticleDOI
TL;DR: In this paper, a latent trait model is presented for the repeated measurement of ability based on a multidimensional conceptualization of the change process, and an example provided that shows good recovery of both item and ability parameters.
Abstract: A latent trait model is presented for the repeated measurement of ability based on a multidimensional conceptualization of the change process. A simplex structure is postulated to link item performance under a given measurement condition or occasion to initial ability and to one or more modifiabilities that represent individual differences in change. Since item discriminations are constrained to be equal within a measurement condition, the model belongs to the family of multidimensional Rasch models. Maximum likelihood estimators of the item parameters and abilities are derived, and an example provided that shows good recovery of both item and ability parameters. Properties of the model are explored, particularly for several classical issues in measuring change.

284 citations


Journal ArticleDOI
TL;DR: In this paper, a method for structural analysis of multivariate data is proposed that combines features of regression analysis and principal component analysis, which is based on the generalized singular value decomposition of a matrix with certain metric matrices.
Abstract: A method for structural analysis of multivariate data is proposed that combines features of regression analysis and principal component analysis. In this method, the original data are first decomposed into several components according to external information. The components are then subjected to principal component analysis to explore structures within the components. It is shown that this requires the generalized singular value decomposition of a matrix with certain metric matrices. The numerical method based on the QR decomposition is described, which simplifies the computation considerably. The proposed method includes a number of interesting special cases, whose relations to existing methods are discussed. Examples are given to demonstrate practical uses of the method.

188 citations


Journal ArticleDOI
TL;DR: A more useable Monte Carlo simulation method for the uniform distribution over this family of all zero-one matrices with given marginals and with a prespecified set of cells with structural zero entries is given.
Abstract: Data in the form of zero-one matrices where conditioning on the marginals is relevant arise in diverse fields such as social networks and ecology; directed graphs constitute an important special case. An algorithm is given for the complete enumeration of the family of all zero-one matrices with given marginals and with a prespecified set of cells with structural zero entries. Complete enumeration is computationally feasible only for relatively small matrices. Therefore, a more useable Monte Carlo simulation method for the uniform distribution over this family is given, based on unequal probability sampling and ratio estimation. This method is applied to testing reciprocity of choices in social networks.

147 citations


Journal ArticleDOI
TL;DR: A number of methods for the analysis of three-way data are described and shown to be variants of principal components analysis (PCA) of the two-way supermatrix in which each twoway slice is "strung out" into a column vector.
Abstract: A number of methods for the analysis of three-way data are described and shown to be variants of principal components analysis (PCA) of the two-way supermatrix in which each two-way slice is “strung out” into a column vector The methods are shown to form a hierarchy such that each method is a constrained variant of its predecessor A strategy is suggested to determine which of the methods yields the most useful description of a given three-way data set

138 citations


Journal ArticleDOI
TL;DR: A computational procedure to evaluate the approximate minimum rank for a covariance matrix yields those proper communalities for which the unexplained common variance is ignored in low-rank factor analysis and permits a numerical determination of the exact minimum rank of a covariances matrix, within limits of computational accuracy.
Abstract: A concept of approximate minimum rank for a covariance matrix is defined, which contains the (exact) minimum rank as a special case. A computational procedure to evaluate the approximate minimum rank is offered. The procedure yields those proper communalities for which the unexplained common variance, ignored in low-rank factor analysis, is minimized. The procedure also permits a numerical determination of the exact minimum rank of a covariance matrix, within limits of computational accuracy. A set of 180 covariance matrices with known or bounded minimum rank was analyzed. The procedure was successful throughout in recovering the desired rank.

130 citations


Journal ArticleDOI
TL;DR: In this article, a logistic regression model is proposed for estimating the relation between a set of manifest predictors and a latent trait assumed to be measured by a set k dichotomous items.
Abstract: A logistic regression model is suggested for estimating the relation between a set of manifest predictors and a latent trait assumed to be measured by a set ofk dichotomous items. Usually the estimated subject parameters of latent trait models are biased, especially for short tests. Therefore, the relation between a latent trait and a set of predictors should not be estimated with a regression model in which the estimated subject parameters are used as a dependent variable. Direct estimation of the relation between the latent trait and one or more independent variables is suggested instead. Estimation methods and test statistics for the Rasch model are discussed and the model is illustrated with simulated and empirical data.

125 citations


Journal ArticleDOI
TL;DR: In this article, a simple structure rotation of a PCAMIX solution based on the rotation of component scores is proposed, which can be viewed as generalizations of simple structure methods for PCA.
Abstract: Several methods have been developed for the analysis of a mixture of qualitative and quantitative variables, and one, called PCAMIX, includes ordinary principal component analysis (PCA) and multiple correspondence analysis (MCA) as special cases. The present paper proposes several techniques for simple structure rotation of a PCAMIX solution based on the rotation of component scores and indicates how these can be viewed as generalizations of the simple structure methods for PCA. In addition, a recently developed technique for the analysis of mixtures of qualitative and quantitative variables, called INDOMIX, is shown to construct component scores (without rotational freedom) maximizing the quartimax criterion over all possible sets of component scores. A numerical example is used to illustrate the implication that when used for qualitative variables, INDOMIX provides axes that discriminate between the observation units better than do those generated from MCA.

107 citations


Journal ArticleDOI
TL;DR: The basic theme of the EM algorithm, to repeatedly use complete-data methods to solve incomplete data problems, is also a theme of several more recent statistical techniques that combine simulation techniques with complete- data methods to attack problems that are difficult or impossible for EM.
Abstract: The basic theme of the EM algorithm, to repeatedly use complete-data methods to solve incomplete data problems, is also a theme of several more recent statistical techniques. These techniques—multiple imputation, data augmentation, stochastic relaxation, and sampling importance resampling—combine simulation techniques with complete-data methods to attack problems that are difficult or impossible for EM.

98 citations


Journal ArticleDOI
TL;DR: In this article, a theory of essential unidimensionality is proposed for sequences of polytomous items satisfying the reasonable assumption that the expected amount of credit awarded increases with examinee ability.
Abstract: A definition ofessential independence is proposed for sequences of polytomous items. For items satisfying the reasonable assumption that the expected amount of credit awarded increases with examinee ability, we develop a theory ofessential unidimensionality which closely parallels that of Stout. Essentially unidimensional item sequences can be shown to have a unique (up to change-of-scale) dominant underlying trait, which can be consistently estimated by a monotone transformation of the sum of the item scores. In more general polytomous-response latent trait models (with or without ordered responses), anM-estimator based upon maximum likelihood may be shown to be consistent for θ under essentially unidimensional violations of local independence and a variety of monotonicity/identifiability conditions. A rigorous proof of this fact is given, and the standard error of the estimator is explored. These results suggest that ability estimation methods that rely on the summation form of the log likelihood under local independence should generally be robust under essential independence, but standard errors may vary greatly from what is usually expected, depending on the degree of departure from local independence. An index of departure from local independence is also proposed.

89 citations


Journal ArticleDOI
TL;DR: In this paper, a maximum likelihood based method for simultaneously performing multidimensional scaling and cluster analysis on two-way dominance or profile data was developed, which utilizes mixtures of multivariate conditional normal distributions to estimate a joint space of stimulus coordinates and k vectors, one for each cluster or group, in aT-dimensional space.
Abstract: This paper develops a maximum likelihood based method for simultaneously performing multidimensional scaling and cluster analysis on two-way dominance or profile data. This MULTICLUS procedure utilizes mixtures of multivariate conditional normal distributions to estimate a joint space of stimulus coordinates andK vectors, one for each cluster or group, in aT-dimensional space. The conditional mixture, maximum likelihood method is introduced together with an E-M algorithm for parameter estimation. A Monte Carlo analysis is presented to investigate the performance of the algorithm as a number of data, parameter, and error factors are experimentally manipulated. Finally, a consumer psychology application is discussed involving consumer expertise/experience with microcomputers.

Journal ArticleDOI
TL;DR: In this paper, the authors describe and analyze the performance of an algorithm for rotating a matrix X such that the column-weighted Procrustes distance to Y is minimized.
Abstract: The Procrustes criterion is a common measure for the distance between two matricesX andY, and can be interpreted as the sum of squares of the Euclidean distances between their respective column vectors. Often a weighted Procrustes criterion, using, for example, a weighted sum of the squared distances between the column vectors, is called for. This paper describes and analyzes the performance of an algorithm for rotating a matrixX such that the column-weighted Procrustes distance toY is minimized. The problem of rotatingX intoY such that an aggregate measure of Tucker's coefficient of congruence is maximized is also discussed.

Journal ArticleDOI
TL;DR: A survey of the current state of multidimensional scaling using the city-block metric is presented in this paper, where substantive and theoretical issues, recent algorithmic developments and their implications for seemingly straightforward analyses, isometries with other metrics, links to graph-theoretic models and future prospects are discussed.
Abstract: A survey of the current state of multidimensional scaling using the city-block metric is presented. Topics include substantive and theoretical issues, recent algorithmic developments and their implications for seemingly straightforward analyses, isometries with other metrics, links to graph-theoretic models, and future prospects.

Journal ArticleDOI
TL;DR: In this paper, the authors generalized Kruskal's results to 2×n×n arrays and showed that 2×2×2 arrays can be diagonalized simultaneously with positive probability.
Abstract: A remarkable difference between the concept of rank for matrices and that for three-way arrays has to do with the occurrence of non-maximal rank. The set ofn×n matrices that have a rank less thann has zero volume. Kruskal pointed out that a 2×2×2 array has rank three or less, and that the subsets of those 2×2×2 arrays for which the rank is two or three both have positive volume. These subsets can be distinguished by the roots of a certain polynomial. The present paper generalizes Kruskal's results to 2×n×n arrays. Incidentally, it is shown that twon ×n matrices can be diagonalized simultaneously with positive probability.

Journal ArticleDOI
TL;DR: This paper showed essential equivalences among several linearly constrained correspondence analysis methods, including Fisher's method of additive scoring, Hayashi's second type of quantification method, ter Braak's canonical correspondence analysis, Nishisato's ANOVA of categorical data, correspondence analysis of manipulated contingency tables, BOCKENholt and Bockenholt's least squares canonical analysis with linear constraints, and van der Heijden and Meijerink's zero average restrictions.
Abstract: This paper shows essential equivalences among several methods of linearly constrained correspondence analysis. They include Fisher's method of additive scoring, Hayashi's second type of quantification method, ter Braak's canonical correspondence analysis, Nishisato's type of quantification method, ter Braak's canonical correspondence analysis, Nishisato's ANOVA of categorical data, correspondence analysis of manipulated contingency tables, Bockenholt and Bockenholt's least squares canonical analysis with linear constraints, and van der Heijden and Meijerink's zero average restrictions. These methods fall into one of two classes of methods corresponding to two alternative ways of imposing linear constraints, the reparametrization method and the null space method. A connection between the two is established through Khatri's lemma.

Journal ArticleDOI
TL;DR: A wide variety of paired comparison, triple comparison, and ranking experiments may be viewed as generalized linear models as discussed by the authors, which can be easily fit by maximum likelihood, using the widely available GLIM computer package.
Abstract: A wide variety of paired comparison, triple comparison, and ranking experiments may be viewed as generalized linear models. These include paired comparison models based on both the Bradley-Terry and Thurstone-Mosteller approaches, as well as extensions of these models that allow for ties, order of presentation effects, and the presence of covariates. Moreover, the triple comparison model of Pendergrass and Bradley, as well as models for complete rankings of more than three items, can also be represented as generalized linear models. All such models can be easily fit by maximum likelihood, using the widely available GLIM computer package. Additionally, GLIM enables the computation of likelihood ratio statistics for testing many hypotheses of interest. Examples are presented that cover a variety of cases, along with their implementation on GLIM.

Journal ArticleDOI
TL;DR: In this article, a conditional maximum likelihood estimation procedure and a likelihood-ratio test of hypotheses within the framework of the linear rating scale model (LRSM) are presented.
Abstract: The polytomous unidimensional Rasch model with equidistant scoring, also known as the rating scale model, is extended in such a way that the item parameters are linearly decomposed into certain basic parameters. The extended model is denoted as the linear rating scale model (LRSM). A conditional maximum likelihood estimation procedure and a likelihood-ratio test of hypotheses within the framework of the LRSM are presented. Since the LRSM is a generalization of both the dichotomous Rasch model and the rating scale model, the present algorithm is suited for conditional maximum likelihood estimation in these submodels as well. The practicality of the conditional method is demonstrated by means of a dichotomous Rasch example with 100 items, of a rating scale example with 30 items and 5 categories, and in the light of an empirical application to the measurement of treatment effects in a clinical study.

Journal ArticleDOI
TL;DR: The CANDECOMP algorithm for the PARAFAC analysis ofn×m×p three-way arrays is adapted to handle arrays in whichn>mp more efficiently, and uses less computation time than the original CANDecomP algorithm.
Abstract: The CANDECOMP algorithm for the PARAFAC analysis ofn×m×p three-way arrays is adapted to handle arrays in whichn>mp more efficiently. For such arrays, the adapted algorithm needs less memory space to store the data during the iterations, and uses less computation time than the original CANDECOMP algorithm. The size of the arrays that can be handled by the new algorithm is in no way limited by the number of observation units (n) in the data.

Journal ArticleDOI
TL;DR: In this paper, a person test of the invariance hypothesis is discussed, which is uniformly most powerful and standardized in the sense that the conditional distribution of the test statistic, given a particular level of ability, does not depend on the absolute value of the examinee's ability parameter.
Abstract: The Rasch model predicts that an individual's ability level is invariant over subtests of the total test, and thus, all subtests measure the same latent trait. A person test of this invariance hypothesis is discussed that is uniformly most powerful and standardized in the sense that the conditional distribution of the test statistic, given a particular level of ability, does not depend on the absolute value of the examinee's ability parameter. The test can be routinely performed by applying a computer program designed by and obtainable from the author. Finally, a suboptimal test is derived that is extremely easy to use, and an overall group test of the invariance hypothesis discussed. All tests considered do not rely on asymptotic approximations; hence, they may be applied when the test is of only moderate length and the group of examinees is small.

Journal ArticleDOI
TL;DR: A reparameterization of a latent class model is presented to simultaneously classify and scale nominal and ordered categorical choice data to generate a graphical, multidimensional representation of the classification results.
Abstract: A reparameterization of a latent class model is presented to simultaneously classify and scale nominal and ordered categorical choice data. Latent class-specific probabilities are constrained to be equal to the preference probabilities from a probabilistic ideal-point or vector model that yields a graphical, multidimensional representation of the classification results. In addition, background variables can be incorporated as an aid to interpreting the latent class-specific response probabilities. The analyses of synthetic and real data sets illustrate the proposed method.

Journal ArticleDOI
TL;DR: In this article, a new stochastic multidimensional scaling procedure for the analysis of three-mode, three-way pick any/J data is presented, which provides either a vector or ideal-point model to represent the structure in such data, as well as floating model specifications (e.g., different vectors or ideal points for different choice settings).
Abstract: This paper presents a new stochastic multidimensional scaling procedure for the analysis of three-mode, three-way pick any/J data. The method provides either a vector or ideal-point model to represent the structure in such data, as well as “floating” model specifications (e.g., different vectors or ideal points for different choice settings), and various reparameterization options that allow the coordinates of ideal points, vectors, or stimuli to be functions of specified background variables. A maximum likelihood procedure is utilized to estimate a joint space of row and column objects, as well as a set of weights depicting the third mode of the data. An algorithm using a conjugate gradient method with automatic restarts is developed to estimate the parameters of the models. A series of Monte Carlo analyses are carried out to investigate the performance of the algorithm under diverse data and model specification conditions, examine the statistical properties of the associated test statistic, and test the robustness of the procedure to departures from the independence assumptions. Finally, a consumer psychology application assessing the impact of situational influences on consumers' choice behavior is discussed.

Journal ArticleDOI
TL;DR: In this paper, necessary and sufficient conditions for the local equivalence of two expanded identified models were given for fitting the more restricted model M�1 and MႷ2 under several regularity conditions.
Abstract: Defining equivalent models as those that reproduce the same set of covariance matrices, necessary and sufficient conditions are stated for the local equivalence of two expanded identified modelsM 1 andM 2 when fitting the more restricted modelM 0. Assuming several regularity conditions, the rank deficiency of the Jacobian matrix, composed of derivatives of the covariance elements with respect to the union of the free parameters ofM 1 andM 2 (which characterizes modelM 12), is a necessary and sufficient condition for the local equivalence ofM 1 andM 2. This condition is satisfied, in practice, when the analysis dealing with the fitting ofM 0, predicts that the decreases in the chi-square goodness-of-fit statistic for the fitting ofM 1 orM 2, orM 12 are all equal for any set of sample data, except on differences due to rounding errors.

Journal ArticleDOI
TL;DR: In this article, the authors show that an unrestricted (or true) model witht parameters should be chosen over a restricted model withm parameters if (Pt2−Pm2)>(1−Pt 2)(t−m)/n, where pt2 and pm2 are the population coefficients of determination of the unrestricted and restricted models, respectively, and n is the sample size.
Abstract: Mean squared error of prediction is used as the criterion for determining which of two multiple regression models (not necessarily nested) is more predictive. We show that an unrestricted (or true) model witht parameters should be chosen over a restricted (or misspecified) model withm parameters if (Pt2−Pm2)>(1−Pt2)(t−m)/n, wherePt2 andPm2 are the population coefficients of determination of the unrestricted and restricted models, respectively, andn is the sample size. The left-hand side of the above inequality represents the squared bias in prediction by using the restricted model, and the right-hand side gives the reduction in variance of prediction error by using the restricted model. Thus, model choice amounts to the classical statistical tradeoff of bias against variance. In practical applications, we recommend thatP2 be estimated by adjustedR2. Our recommendation is equivalent to performing theF-test for model comparison, and using a critical value of 2−(m/n); that is, ifF>2−(m/n), the unrestricted model is recommended; otherwise, the restricted model is recommended.

Journal ArticleDOI
TL;DR: Attention is focused on a particularly general triad and preferential choice model previously requiring the numerical evaluation of a 2n-fold integral, wheren is the number of elements in the vectors representing the psychological magnitudes, and the significance of this form to multidimensional scaling and computational efficiency.
Abstract: Multidimensional probabilistic models of behavior following similarity and choice judgements have proven to be useful in representing multidimensional percepts in Euclidean and non-Euclidean spaces. With few exceptions, these models are generally computationally intense because they often require numerical work with multiple integrals. This paper focuses attention on a particularly general triad and preferential choice model previously requiring the numerical evaluation of a 2n-fold integral, wheren is the number of elements in the vectors representing the psychological magnitudes. Transforming this model to an indefinite quadratic form leads to a single integral. The significance of this form to multidimensional scaling and computational efficiency is discussed.

Journal ArticleDOI
TL;DR: In this article, the item characteristic curve (ICC) is estimated by using polynomial regression splines, which provide a more flexible family of functions than is given by the three-parameter logistic family.
Abstract: The item characteristic curve (ICC), defining the relation between ability and the probability of choosing a particular option for a test item, can be estimated by using polynomial regression splines. These provide a more flexible family of functions than is given by the three-parameter logistic family. The estimation of spline ICCs is described by maximizing the marginal likelihood formed by integrating ability over a beta prior distribution. Some simulation results compare this approach with the joint estimation of ability and item parameters.

Journal ArticleDOI
TL;DR: This paper examined fourteen multiple-choice achievement tests with from 20 to 50 items to see if it was possible for them to produce item response vectors with multiple maxima; such vectors were found for all the tests.
Abstract: Samejima identified the possibility of multiple solutions to the likelihood equation (multiple maxima in the likelihood function) for estimating an examinee's trait value for the three-parameter logistic model. In the practical applications that Lord studied, he found that multiple solutions did not occur when the number of items was ≥20. In the present paper, fourteen multiple-choice achievement tests with from 20 to 50 items were examined to see if it was possible for them to produce item response vectors with multiple maxima; such vectors were found for all the tests. Examination of response vectors for large groups of real examinees found that from 0 to 3.1% of them had response vectors with multiple maxima. The implications of these results for multiple-choice tests are discussed.

Journal ArticleDOI
TL;DR: A generalization of the SMACOF method is proposed to resolve the difficulty that is based on the same rationale frequently involved in robust fitting with least absolute residuals.
Abstract: The usual convergence proof of the SMACOF algorithm model for least squares multidimensional scaling critically depends on the assumption of nonnegativity of the quantities to be fitted, called the pseudodistances. When this assumption is violated, erratic convergence behavior is known to occur. Three types of circumstances in which some of the pseudodistances may become negative are outlined: nonmetric multidimensional scaling with normalization on the variance, metric multidimensional scaling including an additive constant, and multidimensional scaling under the city-block distance model. A generalization of the SMACOF method is proposed to resolve the difficulty that is based on the same rationale frequently involved in robust fitting with least absolute residuals.

Journal ArticleDOI
TL;DR: In this article, the maximum probability of drawing incorrect conclusions about the signs of differences in the presence of disordinal interaction was derived for two-factor designs in which the factor of interest has two levels and the number of levels of the other factor varies.
Abstract: In a factorial design with two or more factors, there is nonzero interaction when the differences among the levels of one factor vary with levels of other factors. The interaction is disordinal or qualitative with respect to a specific factor, sayA, if the difference between at least two levels ofA is positive for some and negative for some levels of the other factors. Using standard methods of analysis, there is a potentially large probability of drawing incorrect conclusions about the signs of differences in the presence of disordinal interaction. The maximum probability of such incorrect conclusions, or directional errors, is derived for two-factor designs in which the factor of interest has two levels and the number of levels of the other factor varies.

Journal ArticleDOI
TL;DR: In this article, it was shown that equivalent coordinate matrices are not granted at global minima when the symmetric matrix is not Gramian, or when these matrices have Gramian solutions but the solution is not globally optimal.
Abstract: Carroll and Chang have claimed that CANDECOMP applied to symmetric matrices yields equivalent coordinate matrices, as needed for INDSCAL. Although this claim has appeared to be valid for all practical purposes, it has gone without a rigorous mathematical footing. The purpose of the present paper is to clarify CANDECOMP in this respect. It is shown that equivalent coordinate matrices are not granted at global minima when the symmetric matrices are not Gramian, or when these matrices are Gramian but the solution not globally optimal.

Journal ArticleDOI
TL;DR: An alternating least squares method for iteratively fitting the longitudinal reduced-rank regression model is proposed in this paper. But the method uses ordinary least squares and majorization substeps to estimate the unknown parameters in the system and measurement equations of the model.
Abstract: An alternating least squares method for iteratively fitting the longitudinal reduced-rank regression model is proposed. The method uses ordinary least squares and majorization substeps to estimate the unknown parameters in the system and measurement equations of the model. In an example with cross-sectional data, it is shown how the results conform closely to results from eigenanalysis. Optimal scaling of nominal and ordinal variables is added in a third substep, and illustrated with two examples involving cross-sectional and longitudinal data.