Title
A Comparison of the LR and DFIT Frameworks of Differential Functioning Applied to the Generalized Graded Unfolding Model
Abbreviated Journal Title
Appl. Psychol. Meas.
Keywords
differential item functioning; measurement invariance; unfolding; item; response theory; likelihood ratio; differential functioning of items and; tests; generalized graded unfolding model; Monte Carlo; LIKELIHOOD RATIO TEST; ITEM RESPONSE THEORY; PARAMETER-ESTIMATION; DOMINANCE MODELS; DIF; TESTS; LINKING; SCALE; STEP; PERSONALITY; Social Sciences, Mathematical Methods; Psychology, Mathematical
Abstract
Recently, applied psychological measurement researchers have become interested in the application of the generalized graded unfolding model (GGUM), a parametric item response theory model that posits an ideal point conception of the relationship between latent attributes and observed item responses. Little attention has been given to considerations for the detection of differential item functioning (DIF) under the GGUM. In this article, the authors present a Monte Carlo simulation meant to assess the efficacy of the likelihood ratio (LR) and differential functioning of items and tests (DFIT) frameworks, two popular ways of detecting DIF. Findings indicate a marked superiority of the LR approach over DFIT in terms of true and false positive rates under the GGUM. The discussion centers on possible explanations for the poor performance of the DFIT framework in detecting DIF under the GGUM and addresses limitations of the current study as well as future research directions.
Journal Title
Applied Psychological Measurement
Volume
35
Issue/Number
8
Publication Date
1-1-2011
Document Type
Article
Language
English
First Page
623
Last Page
642
WOS Identifier
ISSN
0146-6216
Recommended Citation
"A Comparison of the LR and DFIT Frameworks of Differential Functioning Applied to the Generalized Graded Unfolding Model" (2011). Faculty Bibliography 2010s. 1136.
https://stars.library.ucf.edu/facultybib2010/1136
Comments
Authors: contact us about adding a copy of your work at STARS@ucf.edu