Premium
Computerized Adaptive Testing in Early Education: Exploring the Impact of Item Position Effects on Ability Estimation
Author(s) -
Albano Anthony D.,
Cai Liuhan,
Lease Erin M.,
McConnell Scott R.
Publication year - 2019
Publication title -
journal of educational measurement
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.917
H-Index - 47
eISSN - 1745-3984
pISSN - 0022-0655
DOI - 10.1111/jedm.12215
Subject(s) - item response theory , computerized adaptive testing , context (archaeology) , multilevel model , position (finance) , psychology , estimation , relevance (law) , psychometrics , test (biology) , econometrics , applied psychology , statistics , developmental psychology , mathematics , paleontology , management , finance , political science , law , economics , biology
Studies have shown that item difficulty can vary significantly based on the context of an item within a test form. In particular, item position may be associated with practice and fatigue effects that influence item parameter estimation. The purpose of this research was to examine the relevance of item position specifically for assessments used in early education, an area of testing that has received relatively limited psychometric attention. In an initial study, multilevel item response models fit to data from an early literacy measure revealed statistically significant increases in difficulty for items appearing later in a 20‐item form. The estimated linear change in logits for an increase of 1 in position was .024, resulting in a predicted change of .46 logits for a shift from the beginning to the end of the form. A subsequent simulation study examined impacts of item position effects on person ability estimation within computerized adaptive testing. Implications and recommendations for practice are discussed.