2020
DOI: 10.3758/s13428-020-01498-x
|View full text |Cite
|
Sign up to set email alerts
|

Accuracy of performance-test linking based on a many-facet Rasch model

Abstract: Performance assessments, in which human raters assess examinee performance in practical tasks, have attracted much attention in various assessment contexts involving measurement of higher-order abilities. However, difficulty persists in that ability measurement accuracy strongly depends on rater and task characteristics such as rater severity and task difficulty. To resolve this problem, various item response theory (IRT) models incorporating rater and task parameters, including many-facet Rasch models (MFRMs)… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
8
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
3
1

Relationship

1
7

Authors

Journals

citations
Cited by 16 publications
(8 citation statements)
references
References 46 publications
0
8
0
Order By: Relevance
“…1) To appropriately estimate the item parameters based on the original GMFRM while ensuring parameter linking, we require a scored essay dataset in which some examinees answered all the items [83], [84]. However, almost none of the existing datasets that are used for AES studies include such examinees.…”
Section: A Generalized Many-facet Rasch Modelmentioning
confidence: 99%
“…1) To appropriately estimate the item parameters based on the original GMFRM while ensuring parameter linking, we require a scored essay dataset in which some examinees answered all the items [83], [84]. However, almost none of the existing datasets that are used for AES studies include such examinees.…”
Section: A Generalized Many-facet Rasch Modelmentioning
confidence: 99%
“…TAM can estimate student and item measures. The probability that a student will respond to an item correctly was determined by the difference between the student's achievement level and the item's difficulty [15].…”
Section: Measurement Modelmentioning
confidence: 99%
“…Thus, we conducted the same experiment as described above assuming a practice situation where few raters are assigned to each examinee. Concretely, in Procedure 2, we first assigned two raters to each examinee based on a systematic link design (Shin et al 2019;Uto 2020;Wind and Jones 2019), and then we generated the data based on the rater assignment. The examples of a fully crossed design and a systematic link design are illustrated in Tables 6 and 7, where checkmarks indicate an assigned rater, and blank cells indicate that no rater was assigned.…”
Section: Accuracy Of Ability Measurementmentioning
confidence: 99%
“…The written essays were evaluated by 18 raters using a rubric consisting of 9 evaluation items divided into 4 rating categories. We assigned four raters to each essay based on a systematic links design (Shin et al 2019;Uto 2020;Wind and Jones 2019) to reduce the raters' assessment workload. The evaluation items column in Table 9 lists the abstracts of the evaluation items in the rubric, and was created based on two writing assessment rubrics proposed by Matsushita et al (2013), Nakajima (2017 for Japanese university students.…”
Section: Actual Datamentioning
confidence: 99%