Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing 2017
DOI: 10.18653/v1/d17-1116
|View full text |Cite
|
Sign up to set email alerts
|

ConStance: Modeling Annotation Contexts to Improve Stance Classification

Abstract: Manual annotations are a prerequisite for many applications of machine learning. However, weaknesses in the annotation process itself are easy to overlook. In particular, scholars often choose what information to give to annotators without examining these decisions empirically. For subjective tasks such as sentiment analysis, sarcasm, and stance detection, such choices can impact results. Here, for the task of political stance detection on Twitter, we show that providing too little context can result in noisy … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
36
0

Year Published

2018
2018
2022
2022

Publication Types

Select...
6
3

Relationship

1
8

Authors

Journals

citations
Cited by 26 publications
(37 citation statements)
references
References 22 publications
1
36
0
Order By: Relevance
“…Hypothesis: Individual annotators differ in annotation similarity in the contextual presentations, compared to the randomized presentation. Joseph et al in [25] show that while insufficient context results in noisy and uncertain annotations, an overabundance of context may cause the context to outweigh other signals and lead to lower agreement. Further, contextual information biases different people differently on both temporal and intensity metrics [26,27].…”
Section: Questionmentioning
confidence: 99%
“…Hypothesis: Individual annotators differ in annotation similarity in the contextual presentations, compared to the randomized presentation. Joseph et al in [25] show that while insufficient context results in noisy and uncertain annotations, an overabundance of context may cause the context to outweigh other signals and lead to lower agreement. Further, contextual information biases different people differently on both temporal and intensity metrics [26,27].…”
Section: Questionmentioning
confidence: 99%
“…– Manual annotation may yield subjective and noisy labels . Many factors affect the quality of human-annotations, including: (i) unreliable annotators, (ii) poorly specified annotation tasks and guidelines, (iii) poor category design (categories that are too broad, too narrow, or too vague), or (iv) insufficient information to make a reliable assessment (Cheng and Cosley, 2013 ; Joseph et al, 2017 ). Though the goal of an assessment task is to provide human input, underspecification or appeal to subjective judgment can introduce unintended biases that are often hard to detect.…”
Section: Issues Introduced While Processing Datamentioning
confidence: 99%
“…Recent work has shown that considering language within the context of user attributes can improve classification accuracy (Volkova et al, 2013;Bamman et al, 2014;Yang and Eisenstein, 2015;Hovy, 2015;Kulkarni et al, 2016;Lynn et al, 2017). Other work has used network or other meta data, such as in Bamman and Smith (2015); Johnson and Goldwasser (2016); Joseph et al (2017); Khattri et al (2015). In a sense these trail-blazing works might be viewed as case studies on user attributes -identifying particular pieces of information for particular tasks where user information has lead to an advantage.…”
Section: Introductionmentioning
confidence: 99%