2008
DOI: 10.1348/000711006x126600
|View full text |Cite
|
Sign up to set email alerts
|

Computing inter‐rater reliability and its variance in the presence of high agreement

Abstract: Pi (pi) and kappa (kappa) statistics are widely used in the areas of psychiatry and psychological testing to compute the extent of agreement between raters on nominally scaled data. It is a fact that these coefficients occasionally yield unexpected results in situations known as the paradoxes of kappa. This paper explores the origin of these limitations, and introduces an alternative and more stable agreement coefficient referred to as the AC1 coefficient. Also proposed are new variance estimators for the mult… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

7
1,139
1
19

Year Published

2009
2009
2023
2023

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 1,407 publications
(1,248 citation statements)
references
References 15 publications
(24 reference statements)
7
1,139
1
19
Order By: Relevance
“…A majority rules approach was taken by the vEEG reviewers to identify GTCSs, which were required to involve a tonic phase followed immediately by a clonic phase 14. Gwet's AC215 was used to calculate inter‐rater agreement for device placement and categorization of events as GTCSs.…”
Section: Methodsmentioning
confidence: 99%
“…A majority rules approach was taken by the vEEG reviewers to identify GTCSs, which were required to involve a tonic phase followed immediately by a clonic phase 14. Gwet's AC215 was used to calculate inter‐rater agreement for device placement and categorization of events as GTCSs.…”
Section: Methodsmentioning
confidence: 99%
“…If |AVGC-0.5|<sVGC then the difference was considered not significant (n.s.). Inter-observer reliability was assessed for each setting, criterion and type of examination using Gwet's AC1 coefficient (20) . The AC1 coefficient is an alternative to the Kappa coefficient, which is similar in formulation but does not suffer from some of the statistical problems of the Kappa coefficient (21) .…”
Section: Discussionmentioning
confidence: 99%
“…Traditionally, Cohen's Kappa or Scott's Pi has been used to evaluate reliability of human judgment. According to Gwet [13], those ratios can be a problem when the extent of agreement between evaluators is too high. To avoid obtaining biased reliability, we adopted the AC 1 reliability measure [12,13].…”
Section: Effect Of the Constant Comparison Methodsmentioning
confidence: 99%
“…According to Gwet [13], those ratios can be a problem when the extent of agreement between evaluators is too high. To avoid obtaining biased reliability, we adopted the AC 1 reliability measure [12,13]. Overall, the average of AC1 shows positive signs.…”
Section: Effect Of the Constant Comparison Methodsmentioning
confidence: 99%