2024
DOI: 10.1037/aca0000470
|View full text |Cite
|
Sign up to set email alerts
|

Reexamining subjective creativity assessments in science tasks: An application of the rater-mediated assessment framework and many-facet Rasch model.

Abstract: Subjective creativity assessments that are originally developed as the Consensual Assessment Technique (CAT) rely on human raters to score creativity of the products. A few controversial issues in this approach are related to tasks, subjects, raters’ qualifications and performance, and methods for analyzing rating scores. This study addressed these issues under the theoretical framework of rater-mediated assessment and Rasch measurement theory. Data were collected from three groups of raters with different lev… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Publication Types

Select...
6

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(5 citation statements)
references
References 75 publications
(142 reference statements)
0
5
0
Order By: Relevance
“…The application of different criteria seems to be a more serious issue than that of different scoring procedures or scales because consensus can be hardly reached over these criteria even among judges with similar backgrounds (Long, 2014b; Long & Pang, 2015). Therefore, given that variations in rating criteria and procedures are mainly related to raters, creativity researchers must pay more attention to issues such as rater effects, rater cognition, and how they affect the validity of the assessment (Kaufman et al, 2008; Long, 2014b; Long & Pang, 2015; Primi et al, 2019; Wang & Long, in press).…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…The application of different criteria seems to be a more serious issue than that of different scoring procedures or scales because consensus can be hardly reached over these criteria even among judges with similar backgrounds (Long, 2014b; Long & Pang, 2015). Therefore, given that variations in rating criteria and procedures are mainly related to raters, creativity researchers must pay more attention to issues such as rater effects, rater cognition, and how they affect the validity of the assessment (Kaufman et al, 2008; Long, 2014b; Long & Pang, 2015; Primi et al, 2019; Wang & Long, in press).…”
Section: Resultsmentioning
confidence: 99%
“…Primi et al (2019) used Many-Facet Rasch Modeling (MFRM) to model rater effects and investigate the effect of missingness on product-based creativity assessment. New methodological theories and methods have also been proposed, such as Myszkowski and Storme’s (2019) Judge Response Theory and Myszkowski’s (2019) jrt developed in R. Wang and Long (in press) recently examined the rater effects under the theoretical framework of rater-mediated assessment, which refers to “assessments in which raters evaluate test-taker performances and use rating scale categories to describe the level of performance on one or more domains” (Engelhard & Wind, 2019, p. 475).…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…In the field of creativity research, the process of selecting and training judges, enacting the coding scheme, and calculating the reliability of the ratings is most often seen as a means to an end: the main intention in the field is to study originality or creativity itself, and therefore, we aim for reliable codes in order to methodologically support our substantive work. However, a small but notable body of research has focused on the process of human rating eo ipso , empirically investigating and making recommendations on the types of tasks, participants, coders, and trainings that best support rater agreement and reliability (see Wang & Long, 2022 for a recent example). In this paper, we build on that existing literature to develop a model to inform the field's understanding of rater variance when coding the originality of children's responses to creativity assessments.…”
Section: Summary Of Past Work On Judgment Quality In Creativity Asses...mentioning
confidence: 99%
“…In accordance with standard scoring rubrics, 30 ships were randomly selected and scored by two external senior teachers, who had more than fifteen years of teaching experience and participated in relative research more than five years. A consensual assessment technique was used in this research (Wang & Long, 2022; Weiss et al, 2021). The correlation coefficients for design concept variation, novelty, and feasibility were 0.969, 0.853, and 0.698, indicating a high level of reliability (Mugenda & Mugenda, 2003).…”
Section: Research Design and Implementationmentioning
confidence: 99%