Zur Messgüte von geschlossenen und offenen Antwortformaten in Lernstandserhebungen
Psychometric properties of multiple-choice and constructed response formats in proficiency tests
Journal article › Research › Peer reviewed
Publication data
By | Johannes Schult, Marlit Annalena Lindner |
Original language | German |
Published in | Psychologie in Erziehung und Unterricht, 66(4) |
Pages | 260-272 |
Editor (Publisher) | Ernst Reinhardt Verlag |
ISSN | 0342-183X |
DOI/Link | https://doi.org/10.2378/peu2018.art31d, https://www.reinhardt-journals.de/index.php/peu/article/view/7702 |
Publication status | Published – 10.2019 |
Educational large-scale assessments often employ various different response formats. Objective scoring rules are an advantage of multiple-choice (MC) items over constructed response (CR) items. State-wide educational assessments are often coded by the teachers on location rather than centrally by trained raters. The present study investigates format-specific differential validity, using a mathematics competency assessment in the domain of numbers and operations. The analysis is based on a subsample of the “Lernstand 5” pilot study 2016 (n = 1205 fourth-graders). Using MC items and short answer (CR) items respectively, format-specific scales were created with the same number of items per scale and comparable reliabilities. Both format-specific scales showed good criterion validity (rMC = .57; rCR = .60) with the Mathematics grade. The short answer scale’s validity did not differ significantly from the MC scale’s validity (Δr = 0.03, p = .15). These results support the conclusion that teacher-coded assessments with both response formats can yield a valid measurement in educational large-scale assessments.