scispace - formally typeset
Search or ask a question
Author

Fumiko Samejima

Other affiliations: University of New Brunswick
Bio: Fumiko Samejima is an academic researcher from University of Tennessee. The author has contributed to research in topics: Item response theory & Computerized adaptive testing. The author has an hindex of 18, co-authored 25 publications receiving 4724 citations. Previous affiliations of Fumiko Samejima include University of New Brunswick.

Papers
More filters
Journal ArticleDOI
TL;DR: In this article, the authors considered the problem of estimating latent ability using the entire response pattern of free-response items, first in the general case and then in the case where the items are scored in a graded way, especially when the thinking process required for solving each item is assumed to be homogeneous.
Abstract: Estimation of latent ability using the entire response pattern of free-response items is discussed, first in the general case and then in the case where the items are scored in a graded way, especially when the thinking process required for solving each item is assumed to be homogeneous. The maximum likelihood estimator, the Bayes modal estimator, and the Bayes estimator obtained by using the mean-square error multiplied by the density function of the latent variate as the loss function are taken as our estimators. Sufficient conditions for the existence of a unique maximum likelihood estimator and a unique Bayes modal estimator are formulated with respect to an individual item rather than with respect to a whole set of items, which are useful especially in the situation where we are free to choose optimal items for a particular examinee out of the item library in which a sufficient number of items are stored with reliable quality controls. Advantages of the present methods are investigated by comparing them with those which make use of conventional dichotomous items or test scores, theoretically as well as empirically, in terms of the amounts of information, the standard errors of estimators, and the mean-square errors of estimators. The utility of the Bayes modal estimator as a computational compromise for the Bayes estimator is also discussed and observed. The relationship between the formula for the item characteristic function and the philosophy of scoring is observed with respect to dichotomous items.

3,031 citations

Book ChapterDOI
01 Jan 1997
TL;DR: The graded response model as mentioned in this paper is a family of mathematical models that deal with ordered polytomous categories, such as rating such as letter grading, A, B, C, D, and F, used in the evaluation of students' performance; strongly disagree, disagree, agree, and strongly agree used in attitude surveys; or partial credit given in accordance with an examinee's degree of attainment in solving a problem.
Abstract: The graded response model represents a family of mathematical models that deals with ordered polytomous categories. These ordered categories include rating such as letter grading, A, B, C, D, and F, used in the evaluation of students’ performance; strongly disagree, disagree, agree, and strongly agree, used in attitude surveys; or partial credit given in accordance with an examinee’s degree of attainment in solving a problem.

967 citations

Journal Article

232 citations

Journal ArticleDOI
TL;DR: The homogeneous case of the continuous response model is expanded to the multi-dimensional latent space, and the normal ogive model is presented, and it is found that there is a vector of sufficient statistics for estimating the subject's vector of latent traits, given the item parameter vectors.
Abstract: The homogeneous case of the continuous response model is expanded to the multi-dimensional latent space, and the normal ogive model is presented. The operating density characteristic of the continuous item response and the vector of basic functions are developed. It is found out that there is a vector of sufficient statistics for estimating the subject's vector of latent traits, given the item parameter vectors. The relationship between the model and the linear factor analysis is observed. The matrix of item response information functions is introduced. Some additional observations are also made.

128 citations

Journal ArticleDOI
TL;DR: It is emphasized that the standard error of estimation should be considered as the major index of dependability, as opposed to the reliability of a test.
Abstract: Several important and useful implications in latent trait theory, with direct implications for individualized adaptive or tailored testing, are pointed out. A way of using the information function in tailored testing in connection with the standard error of estimation of the ability level using maximum likelihood estimation is suggested. It is emphasized that the standard error of estimation should be considered as the major index of dependability, as opposed to the reliability of a test. The concept of weak parallel forms is expanded to test-

85 citations


Cited by
More filters
Journal ArticleDOI
TL;DR: The brms package implements Bayesian multilevel models in R using the probabilistic programming language Stan, allowing users to fit linear, robust linear, binomial, Poisson, survival, ordinal, zero-inflated, hurdle, and even non-linear models all in a multileVEL context.
Abstract: The brms package implements Bayesian multilevel models in R using the probabilistic programming language Stan. A wide range of distributions and link functions are supported, allowing users to fit - among others - linear, robust linear, binomial, Poisson, survival, ordinal, zero-inflated, hurdle, and even non-linear models all in a multilevel context. Further modeling options include autocorrelation of the response variable, user defined covariance structures, censored data, as well as meta-analytic standard errors. Prior specifications are flexible and explicitly encourage users to apply prior distributions that actually reflect their beliefs. In addition, model fit can easily be assessed and compared with the Watanabe-Akaike information criterion and leave-one-out cross-validation.

4,353 citations

Journal ArticleDOI
TL;DR: In this paper, an unidimensional latent trait model for responses scored in two or more ordered categories is developed, which can be viewed as an extension of Andrich's Rating Scale model to situations in which ordered response alternatives are free to vary in number and structure from item to item.
Abstract: A unidimensional latent trait model for responses scored in two or more ordered categories is developed. This “Partial Credit” model is a member of the family of latent trait models which share the property of parameter separability and so permit “specifically objective” comparisons of persons and items. The model can be viewed as an extension of Andrich's Rating Scale model to situations in which ordered response alternatives are free to vary in number and structure from item to item. The difference between the parameters in this model and the “category boundaries” in Samejima's Graded Response model is demonstrated. An unconditional maximum likelihood procedure for estimating the model parameters is developed.

3,368 citations

Journal ArticleDOI
TL;DR: The authors show how IRT techniques can be used to develop new attachment scales with desirable psychometric properties, and indicate that commonly used attachment scales can be improved in a number of important ways.
Abstract: Self-report measures of adult attachment are typically scored in ways (e.g., averaging or summing items) that can lead to erroneous inferences about important theoretical issues, such as the degree of continuity in attachment security and the differential stability of insecure attachment patterns. To determine whether existing attachment scales suffer from scaling problems, the authors conducted an item response theory (IRT) analysis of 4 commonly used self-report inventories: Experiences in Close Relationships scales (K. A. Brennan, C. L. Clark, & P. R. Shaver, 1998), Adult Attachment Scales (N. L. Collins & S. J. Read, 1990), Relationship Styles Questionnaire (D. W. Griffin & K. Bartholomew, 1994) and J. Simpson's (1990) attachment scales. Data from 1,085 individuals were analyzed using F. Samejima's (1969) graded response model. The authors' findings indicate that commonly used attachment scales can be improved in a number of important ways. Accordingly, the authors show how IRT techniques can be used to develop new attachment scales with desirable psychometric properties.

2,883 citations

Journal ArticleDOI
TL;DR: In this paper, a rating response mechanism for ordered categories, which is related to the traditional threshold formulation but distinctively different from it, is formulated, in which subject and item parameters are derived in terms of thresholds on a latent continuum and discriminations at the thresholds.
Abstract: A rating response mechanism for ordered categories, which is related to the traditional threshold formulation but distinctively different from it, is formulated. In addition to the subject and item parameters two other sets of parameters, which can be interpreted in terms of thresholds on a latent continuum and discriminations at the thresholds, are obtained. These parameters are identified with the category coefficients and the scoring function of the Rasch model for polychotomous responses in which the latent trait is assumed uni-dimensional. In the case where the threshold discriminations are equal, the scoring of successive categories by the familiar assignment of successive integers is justified. In the case where distances between thresholds are also equal, a simple pattern of category coefficients is shown to follow.

2,709 citations