2020 IEEE International Conference on Software Maintenance and Evolution (ICSME) 2020
DOI: 10.1109/icsme46990.2020.00017
|View full text |Cite
|
Sign up to set email alerts
|

Sentiment Analysis for Software Engineering: How Far Can Pre-trained Transformer Models Go?

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

9
80
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
1
1

Relationship

1
6

Authors

Journals

citations
Cited by 80 publications
(89 citation statements)
references
References 35 publications
9
80
0
Order By: Relevance
“…However, not all the solutions support retraining, as in the case of the lexicon-based tools DEVA and SentiStrength-SE. Based on the results of our analysis on the agreement of tools (see Section 5) and in line with previous evidence (Jongeling et al 2017;Zhang et al 2020), we suggest implementing an ensemble of tools with a majority voting system as a possible way to increase the agreement with manual labels when the retraining of the selected solution is not an option.…”
Section: Sentiment Analysis Tools Should Be Retrained If Possible Rather Than Used Off the Shelfsupporting
confidence: 83%
See 1 more Smart Citation
“…However, not all the solutions support retraining, as in the case of the lexicon-based tools DEVA and SentiStrength-SE. Based on the results of our analysis on the agreement of tools (see Section 5) and in line with previous evidence (Jongeling et al 2017;Zhang et al 2020), we suggest implementing an ensemble of tools with a majority voting system as a possible way to increase the agreement with manual labels when the retraining of the selected solution is not an option.…”
Section: Sentiment Analysis Tools Should Be Retrained If Possible Rather Than Used Off the Shelfsupporting
confidence: 83%
“…As for disagreement, the proportion of severe disagreement rates drops to 2% and 3% for GitHub and Stack Overflow, respectively, which is comparable to the disagreement between human raters (see Table 11). More recently, similar findings were presented by Zhang et al (2020). In their study leveraging deep learning for sentiment analysis, they provided evidence on how composition of different classifiers may boost performance.…”
Section: Follow-up Analysis On Majority Votingsupporting
confidence: 65%
“…For each API review aspect, we evaluate the performance in terms of five evaluation metrics (i.e., P, R, F 1, M CC, and AU C) as introduced in Section III-E. RQ1 Can pre-trained transformer-based models achieve better performance than the state-of-the-art approach which is based on traditional machine learning models? Motivation Previous studies have shown the great potential of pre-trained transformer-based models on many software engineering tasks, e.g., sentiment analysis for software data [9] and code summarization [12]. However, the efficacy of the pre-trained transformer-based models for various types of For the summative result in Table IV, we calculate the arithmetic average of the used evaluation metrics of each approach across all the aspects as avg.…”
Section: Discussionmentioning
confidence: 99%
“…Pre-trained Transformer-based Approaches In this study, we consider four popular and state-of-the-art pre-trained transformer-based models which have been utilized in many other software tasks [9], [24], [25], including BERT, RoBERTa, ALBERT, XLNet. We also apply two PTM variants: BERTOverflow [18] that is pre-trained with software engineer in-domain data; CostSensBERT [19] that designed to handle imbalanced data.…”
Section: Implementationsmentioning
confidence: 99%
See 1 more Smart Citation