1981
DOI: 10.1177/001316448104100307
|View full text |Cite
|
Sign up to set email alerts
|

Coefficient Kappa: Some Uses, Misuses, and Alternatives

Abstract: This paper considers some appropriate and inappropriate uses of coefficient kappa and alternative kappa-like statistics. Discussion is restricted to the descriptive characteristics of these statistics for measuring agreement with categorical data in studies of reliability and validity. Special consideration is given to assumptions about whether marginals are fixed a priori, or free to vary. In reliability studies, when marginals are fixed, coefficient kappa is found to be appropriate. When either or both of th… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

7
684
0
20

Year Published

1996
1996
2012
2012

Publication Types

Select...
7
3

Relationship

0
10

Authors

Journals

citations
Cited by 1,170 publications
(762 citation statements)
references
References 13 publications
7
684
0
20
Order By: Relevance
“…27 Within-group differences from pretest (baseline) to posttest (outcome) were analyzed using paired t-tests.…”
Section: Resultsmentioning
confidence: 99%
“…27 Within-group differences from pretest (baseline) to posttest (outcome) were analyzed using paired t-tests.…”
Section: Resultsmentioning
confidence: 99%
“…The kappa coefficient (Cohen, 1960;Fleiss, 1975;Kraemer, 1979;Brennan and Prediger, 1981;Zwick 1988;Warrens, 2008aWarrens, ,b, 2010) is a popular descriptive statistic for summarizing the cross classification of two nominal variables with k ∈ N ≥2 identical categories. These k × k tables occur in various fields of science, including psychometrics, educational measurement, epidemiology, map comparison (Visser and De Nijs, 2006) and content analysis (Krippendorff, 2004).…”
Section: Introductionmentioning
confidence: 99%
“…The κ coefficient (Cohen, 1960;Fleiss, 1981;Brennan & Prediger, 1981;Maclure & Willett, 1987;Zwick, 1988;Hsu & Field, 2003;Warrens, 2008aWarrens, ,b,c, 2010aWarrens, ,b,d, 2011) is a popular descriptive statistic for summarizing the cross classification of two nominal variables with n ∈ N ≥2 identical categories. Originally proposed as a measure of agreement between two observers who each rate the same sample of objects (individuals, observations) on a nominal (unordered) scale with the same number of n categories, κ has been applied to square cross-classifications encountered in psychometrics, educational measurement, epidemiology, diagnostic imaging (Kundel & Polansky, 2003), map comparison (Visser & Nijs, 2006) and content analysis (Krippendorff, 2004).…”
Section: Cohen's Kappamentioning
confidence: 99%