Proceedings of the 3rd International Workshop on Emotion Awareness in Software Engineering 2018
DOI: 10.1145/3194932.3194938
|View full text |Cite
|
Sign up to set email alerts
|

Sentiment and politeness analysis tools on developer discussions are unreliable, but so are people

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 24 publications
(16 citation statements)
references
References 34 publications
0
16
0
Order By: Relevance
“…Therefore, a possible explanation for the low agreement observed is that the benchmarked tools have been originally validated and tuned on gold standards that include manual annotation following different guidelines. As already pointed out by previous research, sentiment annotation is a subjective task, thus even humans might disagree with each-others (Imtiaz et al 2018) if model-driven annotation is not adopted (Novielli et al 2018b). Moreover, Islam and Zibran (2018a) showed how tools exhibit their best performance on the dataset they were originally tested at the time of their release, whereas a drop in performance is observed when they are assessed on a different dataset.…”
Section: Sentiment Analysis Tools Should Be Retrained If Possible Rather Than Used Off the Shelfmentioning
confidence: 97%
“…Therefore, a possible explanation for the low agreement observed is that the benchmarked tools have been originally validated and tuned on gold standards that include manual annotation following different guidelines. As already pointed out by previous research, sentiment annotation is a subjective task, thus even humans might disagree with each-others (Imtiaz et al 2018) if model-driven annotation is not adopted (Novielli et al 2018b). Moreover, Islam and Zibran (2018a) showed how tools exhibit their best performance on the dataset they were originally tested at the time of their release, whereas a drop in performance is observed when they are assessed on a different dataset.…”
Section: Sentiment Analysis Tools Should Be Retrained If Possible Rather Than Used Off the Shelfmentioning
confidence: 97%
“…Of the 𝑛 = 80 papers, 26 mentioned the problem of the lack or scarcity of adaptations of existing sentiment analysis tools to the domain of SE (e.g. [3,8,10,15,27]). However, other problems like sarcasm/irony handling (11) or the subjectivity of manual labeling of data (10) were also mentioned.…”
Section: Difficultiesmentioning
confidence: 99%
“…The authors often stated that existing, domain independent tools lead to poor results in the SE domain (e.g. [8,27,42]). This is because certain terms are used differently in the SE domain than in the nontechnical context, resulting in different sentiments.…”
Section: Rqmentioning
confidence: 99%
“…[24,25,27,30]. In consideration of sentiment analysis for software engineering domain, [8,9,11], mainly focus on deriving developers' opinions/emotions, along with the context. It is observed that the existing tools make dataset driven predictions, where each prediction conflicts with one another [21].…”
Section: Related Workmentioning
confidence: 99%