z-logo
Premium
Exploring Proficiency‐Based vs. Performance‐Based Items With Elicited Imitation Assessment
Author(s) -
Cox Troy L.,
Bown Jennifer,
Burdis Jacob
Publication year - 2015
Publication title -
foreign language annals
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.258
H-Index - 49
eISSN - 1944-9720
pISSN - 0015-718X
DOI - 10.1111/flan.12152
Subject(s) - language proficiency , psychology , imitation , test (biology) , item response theory , test score , standardized test , psychometrics , developmental psychology , mathematics education , social psychology , paleontology , biology
This study investigates the effect of proficiency‐ vs. performance‐based elicited imitation (EI) assessment. EI requires test‐takers to repeat sentences in the target language. The accuracy at which test‐takers are able to repeat sentences highly correlates with test‐takers' language proficiency. However, in EI, the factors that render an item more complex are still being investigated. In order to investigate whether item difficulty and test performance were different between proficiency‐ and performance‐based tests, two EI instruments were created—one to measure proficiency with items from a general corpus and another to measure language for specific purposes (LSP) performance with items from a domain‐specific corpus. The two instruments were then administered to 98 subjects of varying proficiency. The mean score for the LSP performance test ( x ¯  = 0.51) was significantly higher than the mean score for the proficiency test ( x ¯  = 0.44, p < 0.001). In addition, item difficulties for the LSP items were significantly lower than item difficulties for the general items (p < 0.05), indicating that the content of the EI items affected item difficulty. Data suggest that the two approaches to EI assess different constructs and cannot be used interchangeably.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here