Handbook of Linguistic Annotation 2017
DOI: 10.1007/978-94-024-0881-2_11
|View full text |Cite
|
Sign up to set email alerts
|

Inter-annotator Agreement

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
61
0
5

Year Published

2019
2019
2023
2023

Publication Types

Select...
4
3
2

Relationship

0
9

Authors

Journals

citations
Cited by 96 publications
(66 citation statements)
references
References 23 publications
0
61
0
5
Order By: Relevance
“…Point 2: Reporting IAA studies With regard to point 2, much has been said in previous works. Because presenting a detailed report of those works is beyond the scope of this paper, we refer to Krippendorff (1980), Lombard et al (2002), Artstein and Poesio (2008), LeBreton and Senter (2008), Kottner et al (2011) and Artstein (2017), where guidelines and good practice descriptions for applying IAA have been developed. Based on our research, the following shortcomings have been identified.…”
Section: Years Of Iaa In Evaluation Of Nlg Systemsmentioning
confidence: 99%
See 1 more Smart Citation
“…Point 2: Reporting IAA studies With regard to point 2, much has been said in previous works. Because presenting a detailed report of those works is beyond the scope of this paper, we refer to Krippendorff (1980), Lombard et al (2002), Artstein and Poesio (2008), LeBreton and Senter (2008), Kottner et al (2011) and Artstein (2017), where guidelines and good practice descriptions for applying IAA have been developed. Based on our research, the following shortcomings have been identified.…”
Section: Years Of Iaa In Evaluation Of Nlg Systemsmentioning
confidence: 99%
“…The reduction of a statistical test interpretation to a simple number, whilst common, can be arbitrary and accordingly give us little information 8 . For example, Artstein (2017) show that a single label is not sufficient to give a deep understanding of the reliability of an annotation. In this paper, we do not face the problem of how to interpret IAA, rather, we try to tackle the prob-lem of data reliability by suggesting that correlation coefficients and agreement coefficients should be used together to obtain a better assessment of the evaluation data reliability.…”
Section: Years Of Iaa In Evaluation Of Nlg Systemsmentioning
confidence: 99%
“…Our first attempt is to maximize the mutual information between the predictions of p and q. Intuitively, this encourages p and q to agree on some annotation scheme (up to a permutation of labels), modeling the dynamics of inter-annotator agreement (Artstein, 2017). It can be seen as a differentiable generalization of the Brown clustering objective.…”
Section: Generalized Brown Objectivementioning
confidence: 99%
“…The annotation was performed at sentence level and the prescribed labels were accepted using inter-annotator agreement. Inter-annotator agreement is a measure of how well two (or more) annotators make the same annotation decision for a certain label in the entire corpus [39]. We measured the interannotation agreement of the three annotators using Cohen's kappa coefficient [40] and found it substantial (kappa = 0.701) for further analysis and improvement in our case study.…”
Section: B Corpus Annotationmentioning
confidence: 93%