2003
DOI: 10.2307/3345847
|View full text |Cite
|
Sign up to set email alerts
|

Faculty Interjudge Reliability of Music Performance Evaluation

Abstract: Assessment of music performance in authentic contexts remains an underinvestigated area of research. This study is an examination of one such context, the inter-judge reliability of faculty evaluation of end-of-semester applied music performances. Brass (n = 4), percussion (n = 2), woodwind (n = 5), voice (n = 5), piano (n = 3), and string (n = 5) instructors evaluating a recent semester's applied music juries at a large university participated in the study. Each evaluator completed a criterion-specific rating… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

4
77
0
31

Year Published

2005
2005
2014
2014

Publication Types

Select...
7
2
1

Relationship

0
10

Authors

Journals

citations
Cited by 64 publications
(113 citation statements)
references
References 11 publications
(17 reference statements)
4
77
0
31
Order By: Relevance
“…Research in musical adjudication has spawned a number of facetfactorial studies, elegantly summarized in Bergee (2003), as well as additional research. Studies have almost always dealt with intra-and interjudge reliability and issues of validity, as well as with correlations between test items.…”
Section: Jrme 163mentioning
confidence: 99%
“…Research in musical adjudication has spawned a number of facetfactorial studies, elegantly summarized in Bergee (2003), as well as additional research. Studies have almost always dealt with intra-and interjudge reliability and issues of validity, as well as with correlations between test items.…”
Section: Jrme 163mentioning
confidence: 99%
“…With the general consensus on the importance of sound in the domain of music, as "an art of sound" (40), it follows that experts and key decision makers would privilege auditory-related rating in professional evaluation and assessment, even when such items show insufficient reliability (41)(42)(43)(44)(45). However, despite all that is invested in the auditory domain, low interrater correlations suggest that such basis of evaluation is an unreliable process.…”
mentioning
confidence: 99%
“…Each dimension on Form B was rated lower than its counterpart on Form A (since the number &dquo;1&dquo; is considered the &dquo;best&dquo; score, lower scores or ratings are indicated by higher numbers). Paired-samples t-tests revealed significant differences between forms at the .05 level or lower in the following dimensions: tone ( t = -2.27, p = .027), diction ( t = -2.40, p=.02), blend ( t = -3.36, p = .001 ) , intonation ( t = -2.34, p = .023), rhythm ( t = -2.80, p = .007), balance ( = -4.09, p < .001 ) , total score ( = -3.94, p < .001), and rating ( (Garman, Barry, & DeCarbo, 1991;Bergee, 1988Bergee, , 1989Bergee, , 1993Bergee, , 1997Bergee, , 2003 The additional analysis of the means of dimensions, total score, and overall ratings corroborates the above comments (see Table 1 and above t-test results). Form B yielded significantly different ratings in every dimension except interpretation, suggesting that the adjudicators in this setting rated the choirs more severely when using Form B.…”
mentioning
confidence: 99%