Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics 2023
DOI: 10.18653/v1/2023.eacl-main.178
|View full text |Cite
|
Sign up to set email alerts
|

Why Don’t You Do It Right? Analysing Annotators’ Disagreement in Subjective Tasks

Marta Sandri,
Elisa Leonardelli,
Sara Tonelli
et al.

Abstract: Annotators' disagreement in linguistic data has been recently the focus of multiple initiatives aimed at raising awareness on issues related to 'majority voting' when aggregating diverging annotations. Disagreement can indeed reflect different aspects of linguistic annotation, from annotators' subjectivity to sloppiness or lack of enough context to interpret a text.In this work we first propose a taxonomy of possible reasons leading to annotators' disagreement in subjective tasks. Then, we manually label part … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
0
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1
1

Relationship

0
3

Authors

Journals

citations
Cited by 3 publications
(1 citation statement)
references
References 38 publications
0
0
0
Order By: Relevance
“…Other domain tasks are transferable to NLI. Our work can be expanded to test LLMs on other NLP applications (Plank, 2022) such as Question Answering (De Marneffe et al, 2019), Fact Verification (Thorne et al, 2018), and Toxic Language Detection (Schmidt and Wiegand, 2017;Sandri et al, 2023). Further, our method can be applied for tasks that contain disagreements since they are easily transferable to NLI tasks (Dagan et al, 2006) like the QNLI dataset from Table 2, for example, instead of directly asking controversial questions (e.g., abortion) to the model (Santurkar et al, 2023), the question format can be modified into a declarative statement in the premise and place a possible answer in the hypothesis with a binary True/False label (Dagan et al, 2006).…”
Section: Discussionmentioning
confidence: 99%
“…Other domain tasks are transferable to NLI. Our work can be expanded to test LLMs on other NLP applications (Plank, 2022) such as Question Answering (De Marneffe et al, 2019), Fact Verification (Thorne et al, 2018), and Toxic Language Detection (Schmidt and Wiegand, 2017;Sandri et al, 2023). Further, our method can be applied for tasks that contain disagreements since they are easily transferable to NLI tasks (Dagan et al, 2006) like the QNLI dataset from Table 2, for example, instead of directly asking controversial questions (e.g., abortion) to the model (Santurkar et al, 2023), the question format can be modified into a declarative statement in the premise and place a possible answer in the hypothesis with a binary True/False label (Dagan et al, 2006).…”
Section: Discussionmentioning
confidence: 99%