z-logo
Premium
What Do School‐Level Scores From Large‐Scale Assessments Really Measure?
Author(s) -
Sicoly Fiore
Publication year - 2002
Publication title -
educational measurement: issues and practice
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.158
H-Index - 52
eISSN - 1745-3992
pISSN - 0731-1745
DOI - 10.1111/j.1745-3992.2002.tb00103.x
Subject(s) - reading (process) , scale (ratio) , variance (accounting) , psychology , cognition , measure (data warehouse) , subject (documents) , illusion , mathematics education , cognitive psychology , computer science , linguistics , data mining , philosophy , physics , accounting , quantum mechanics , neuroscience , library science , business
Although assessments of mathematics, reading, and writing are assumed to measure distinct academic skills, this may be difficult owing to the pervasive influence of general ability on performance. Factor analyses of school‐level data from 14 large‐scale assessment programs revealed that 80% of the variance in mathematics, reading, and writing scores was due to a common, underlying factor. Multiple regression analyses confirmed that scores contribute little information that is unique to a particular subject (6% or less). Although different assessments may create the illusion of providing unique information, they may be tapping into generic cognitive abilities that cut across content areas. These results raise suspicions about the value and validity of interpretations based on school‐level subject area scores.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here