scispace - formally typeset
Journal ArticleDOI

Taking an Item‐Level Approach to Measuring Change With the Force and Motion Conceptual Evaluation: An Application of Item Response Theory

Robert M. Talbot
- 01 Nov 2013 - 
- Vol. 113, Iss: 7, pp 356-365
TLDR
In this paper, the authors argue that measuring change at the item level (rather than the person level) on the Force and Motion Conceptual Evaluation (FMCE) can provide a more detailed insight into the observed change in students' Newtonian thinking.
Abstract
In order to evaluate the effectiveness of curricular or instructional innovations, researchers often attempt to measure change in students' conceptual understanding of the target subject matter. The measurement of change is therefore a critical endeavor. Often, this is accomplished through pre–post testing using an assessment such as a concept inventory, and aggregate test scores are compared from pre to post-test in order to characterize gains. These comparisons of raw or normalized scores are most often made under the assumptions of Classical Test Theory (CTT). This study argues that measuring change at the item level (rather than the person level) on the Force and Motion Conceptual Evaluation (FMCE) can provide a more detailed insight into the observed change in students' Newtonian thinking. Further, such an approach is more warranted under the assumptions of Item Response Theory (IRT). In comparing item-level measures of change under CTT and IRT measurement models, it was found that the inferences drawn from each analysis are similar, but those derived from IRT modeling stand on a stronger foundation statistically. Second, the IRT approach leads to analyzing common item groupings which provide further information about change at the item and topic level.

read more

Citations
More filters
Journal ArticleDOI

Comparison of Normalized Gain and Cohen's "d" for Analyzing Gains on Concept Inventories.

TL;DR: Normalized gain, commonly used to measure instructional efficacy on conceptual instruments such as the Force Concept Inventory, should be used with care as it is biased in favor of populations with higher pretest scores as mentioned in this paper.
Journal ArticleDOI

A comparison of Hake's g and Cohen's d for analyzing gains on concept inventories

TL;DR: In this paper, the authors compared two commonly used methods for analyzing shifts in concept inventories: Hake's gain (g) and Cohen's d. The results showed that the two methods led to different inferences about student learning and equity due to g being biased in favor of high pretest populations.
Journal ArticleDOI

A scoping review of literature assessing the impact of the learning assistant model

TL;DR: The Learning Assistant (LA) model as mentioned in this paper is a near-peer teaching model designed to facilitate evidenced-based teaching in Science, Technology, Engineering, and Mathematics classrooms, and it has been widely studied.
Journal ArticleDOI

Item-level gender fairness in the Force and Motion Conceptual Evaluation and the Conceptual Survey of Electricity and Magnetism

TL;DR: An item-level analysis of the Force and Motion Conceptual Evaluation and Conceptual Survey of Electricity and Magnetism instruments shows that there are relatively few items that show gender unfairness.
Journal ArticleDOI

Partitioning the Gender Gap in Physics Conceptual Inventories: Force Concept Inventory, Force and Motion Conceptual Evaluation, and Conceptual Survey of Electricity and Magnetism.

TL;DR: The authors found that very little of the gender gap on multiple choice conceptual inventories can be explained by differences in academic performance between men and women, and that academic performance was not correlated with gender stereotypes.
References
More filters
Journal ArticleDOI

Interactive-engagement versus traditional methods: A six-thousand-student survey of mechanics test data for introductory physics courses

TL;DR: In this paper, a survey of pre/post test data using the Halloun-Hestenes Mechanics Diagnostic test or more recent Force Concept Inventory is reported for 62 introductory physics courses enrolling a total number of students.
Book

Applying the Rasch Model: Fundamental Measurement in the Human Sciences

TL;DR: This volume contends that Rasch measurement is the model of choice because it is the closest to realizing the sort of objective fundamental measurement so long revered in the physical sciences.
Journal ArticleDOI

A Rasch Model for Partial Credit Scoring.

TL;DR: In this paper, an unidimensional latent trait model for responses scored in two or more ordered categories is developed, which can be viewed as an extension of Andrich's Rating Scale model to situations in which ordered response alternatives are free to vary in number and structure from item to item.
Related Papers (5)