Zur Messgüte von geschlossenen und offenen Antwortformaten in Lernstandserhebungen

Psychometric properties of multiple-choice and constructed response formats in proficiency tests

Journal articleResearchPeer reviewed

Publication data


ByJohannes Schult, Marlit Annalena Lindner
Original languageGerman
Published inPsychologie in Erziehung und Unterricht, 66(4)
Pages260-272
Editor (Publisher)Ernst Reinhardt Verlag
ISSN0342-183X
DOI/Linkhttps://doi.org/10.2378/peu2018.art31d, https://www.reinhardt-journals.de/index.php/peu/article/view/7702
Publication statusPublished – 10.2019

Educational large-scale assessments often employ various different response formats. Objective scoring rules are an advantage of multiple-choice (MC) items over constructed response (CR) items. State-wide educational assessments are often coded by the teachers on location rather than centrally by trained raters. The present study investigates format-specific differential validity, using a mathematics competency assessment in the domain of numbers and operations. The analysis is based on a subsample of the “Lernstand 5” pilot study 2016 (n = 1205 fourth-graders). Using MC items and short answer (CR) items respectively, format-specific scales were created with the same number of items per scale and comparable reliabilities. Both format-specific scales showed good criterion validity (rMC = .57; rCR = .60) with the Mathematics grade. The short answer scale’s validity did not differ significantly from the MC scale’s validity (Δr = 0.03, p = .15). These results support the conclusion that teacher-coded assessments with both response formats can yield a valid measurement in educational large-scale assessments.