Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.316
|View full text |Cite
|
Sign up to set email alerts
|

Improving Human-Labeled Data through Dynamic Automatic Conflict Resolution

Abstract: This paper develops and implements a scalable methodology for (a) estimating the noisiness of labels produced by a typical crowdsourcing semantic annotation task, and (b) reducing the resulting error of the labeling process by as much as 20-30% in comparison to other common labeling strategies. Importantly, this new approach to the labeling process, which we name Dynamic Automatic Conflict Resolution (DACR), does not require a ground truth dataset and is instead based on inter-project annotation inconsistencie… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
5
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 9 publications
(5 citation statements)
references
References 23 publications
0
5
0
Order By: Relevance
“…In [26], several important aspects of data labeling are discussed. The debate surrounding expert annotators versus non-expert annotators is highlighted, with arguments for the effectiveness of both approaches.…”
Section: Relevant Researchmentioning
confidence: 99%
“…In [26], several important aspects of data labeling are discussed. The debate surrounding expert annotators versus non-expert annotators is highlighted, with arguments for the effectiveness of both approaches.…”
Section: Relevant Researchmentioning
confidence: 99%
“…Finally, there has been considerable work on measuring and rectifying inaccuracies in human annotation (Sun et al, 2020;Wei and Jia, 2021;Gladkoff et al, 2021;Paun et al, 2018). We sidestep this issue by aiming to predict the performance of a single human rater, assuming that if this can be done accurately, conflicts among raters can be resolved in a post-processing step.…”
Section: Chaganty Et Al (2018) Pioneered Control Variatesmentioning
confidence: 99%
“…Collecting ground truth data ("gold" datasets) is time consuming and expensive, and sometimes in-volves heavy engineering efforts (Sun et al, 2020). The confidence score generated by our model offers the potential to perform large-scale evaluations of annotation tasks.…”
Section: Introductionmentioning
confidence: 99%