2019
DOI: 10.1111/jedm.12248
|View full text |Cite
|
Sign up to set email alerts
|

IRT Approaches to Modeling Scores on Mixed‐Format Tests

Abstract: This article considers psychometric properties of composite raw scores and transformed scale scores on mixed‐format tests that consist of a mixture of multiple‐choice and free‐response items. Test scores on several mixed‐format tests are evaluated with respect to conditional and overall standard errors of measurement, score reliability, and classification consistency and accuracy under three item response theory (IRT) frameworks: unidimensional IRT (UIRT), simple structure multidimensional IRT (SS‐MIRT), and b… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

1
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
1

Relationship

2
5

Authors

Journals

citations
Cited by 8 publications
(9 citation statements)
references
References 43 publications
1
8
0
Order By: Relevance
“…The findings showed that the bi-factor MIRT model offered a better solution for Grade 1 in terms of model fit as indicated by the lower −2LL value, and model parsimony as indicated by the lower AIC, AICc, BIC, and SABIC values. This result is consistent with previous bi-factor MIRT applications in language assessment (Cai & Kunnan, 2018; Lee et al, 2019; Min & He, 2014), which also reported relatively better model–data fit of the bi-factor MIRT model over unidimensional IRT models and/or multidimensional generalizations of unidimensional IRT models. However, for Grades 2–12, the fit indices and model parsimony indicators tended to lead to different conclusions.…”
Section: Discussionsupporting
confidence: 92%
“…The findings showed that the bi-factor MIRT model offered a better solution for Grade 1 in terms of model fit as indicated by the lower −2LL value, and model parsimony as indicated by the lower AIC, AICc, BIC, and SABIC values. This result is consistent with previous bi-factor MIRT applications in language assessment (Cai & Kunnan, 2018; Lee et al, 2019; Min & He, 2014), which also reported relatively better model–data fit of the bi-factor MIRT model over unidimensional IRT models and/or multidimensional generalizations of unidimensional IRT models. However, for Grades 2–12, the fit indices and model parsimony indicators tended to lead to different conclusions.…”
Section: Discussionsupporting
confidence: 92%
“…First, as the proposed methods were developed and applied for dichotomous items only, future research would involve applying them to cases with polytomous or mixed‐format items. These methods could also be extended to more complex IRT models such as multidimensional models (e.g., Kim et al., 2020; Lee et al., 2020). Second, a future simulation study would involve study factors that were not considered in the present study, such as extremely small samples (e.g., Kolen, 2020; Peabody, 2020), vertical scaling, and so on.…”
Section: Summary and Discussionmentioning
confidence: 99%
“…Unlike UNT, true‐score equating under the SS‐MIRT framework also requires a multivariate true‐score distribution (see Equation ). Three potential methods appeared in the literature (Lee et al., 2020) to approximate the ability distribution of θ : (a) a quadrature distribution (D‐method), (b) Monte‐Carlo simulation (M‐method), and (c) individual latent‐trait estimates bold-italicθ̂$\widehat{\bm{\theta}}$ (P‐method). In Lee et al.…”
Section: Real Data Illustrationmentioning
confidence: 99%
“…In Lee et al. (2020), these three methods were discussed as a means to analyze psychometrics properties using MIRT models. However, they have not been used or introduced in the equating literature yet.…”
Section: Real Data Illustrationmentioning
confidence: 99%