2021
DOI: 10.3390/ejihpe11040117
|View full text |Cite
|
Sign up to set email alerts
|

On the Treatment of Missing Item Responses in Educational Large-Scale Assessment Data: An Illustrative Simulation Study and a Case Study Using PISA 2018 Mathematics Data

Abstract: Missing item responses are prevalent in educational large-scale assessment studies such as the programme for international student assessment (PISA). The current operational practice scores missing item responses as wrong, but several psychometricians have advocated for a model-based treatment based on latent ignorability assumption. In this approach, item responses and response indicators are jointly modeled conditional on a latent ability and a latent response propensity variable. Alternatively, imputation-b… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
21
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
8

Relationship

3
5

Authors

Journals

citations
Cited by 19 publications
(39 citation statements)
references
References 178 publications
1
21
0
Order By: Relevance
“…The GHP can be seen as a normalized variant of the AIC. A difference in GHP larger than 0.001 is a notable difference regarding global model fit [ 72 , 73 ].…”
Section: Model Selection and Model Uncertaintymentioning
confidence: 99%
“…The GHP can be seen as a normalized variant of the AIC. A difference in GHP larger than 0.001 is a notable difference regarding global model fit [ 72 , 73 ].…”
Section: Model Selection and Model Uncertaintymentioning
confidence: 99%
“…Fourth, we focused on commonly used imputation methods, but these methods are not designed to be used when data are missing not at random. Methods for the imputation of data missing not at random exist and the implementation of these methods may serve to further reduce the observed error due to missing data [ 57 , 58 ]. Future work should focus on understanding how the use of these methods affects DIF estimation error due to missing data.…”
Section: Discussionmentioning
confidence: 99%
“…In LSA studies, students often do not respond to administered items [ 81 , 82 , 83 , 84 , 85 , 86 , 87 ]. Two different types of missing item responses can be distinguished [ 88 ]. First, not reached items [ 89 ] are missing item responses at the end of a test booklet (or an item cluster).…”
Section: Methodsmentioning
confidence: 99%
“…Since PISA 2015, not reached items are treated as non-administered items (i.e., treating it as “NA” in the scaling model), while omitted items are scored as incorrect. Several psychometricians argue that missing item responses should never be scored as incorrect [ 33 , 90 , 91 , 92 , 93 , 94 , 95 , 96 ], while others argue that the treatment of missing item responses is not an empirical question because it should be framed as an issue in scoring, not an issue of missing data modeling [ 45 , 88 , 97 , 98 ].…”
Section: Methodsmentioning
confidence: 99%