2019
DOI: 10.1007/s10664-019-09775-w
|View full text |Cite
|
Sign up to set email alerts
|

SIEVE: Helping developers sift wheat from chaff via cross-platform analysis

Abstract: Software developers have benefited from various sources of knowledge such as forums, question-and-answer sites, and social media platforms to help them in various tasks. Extracting software-related knowledge from different platforms involves many challenges. In this paper, we propose an approach to improve the effectiveness of knowledge extraction tasks by performing crossplatform analysis. Our approach is based on transfer representation learning and word embeddings, leveraging information extracted from a so… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
3
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(4 citation statements)
references
References 86 publications
1
3
0
Order By: Relevance
“…There are no statistically significant differences among FastText and Word2vec models in terms of MCC. Our results confirm previous work (Sulistya et al 2020 ; Mikolov et al 2013 ) which assessed the superiority of Word2vec and FastText in a different context (i.e., text mining). In addition, our work agrees with the findings of previous work (Lau and Baldwin 2016 ) suggesting that Doc2vec creates document embeddings which align with lower frequency words when the documents are short and the corpus is relatively small.…”
Section: Results Of the Empirical Studysupporting
confidence: 92%
See 3 more Smart Citations
“…There are no statistically significant differences among FastText and Word2vec models in terms of MCC. Our results confirm previous work (Sulistya et al 2020 ; Mikolov et al 2013 ) which assessed the superiority of Word2vec and FastText in a different context (i.e., text mining). In addition, our work agrees with the findings of previous work (Lau and Baldwin 2016 ) suggesting that Doc2vec creates document embeddings which align with lower frequency words when the documents are short and the corpus is relatively small.…”
Section: Results Of the Empirical Studysupporting
confidence: 92%
“…In a recent study, Sulistya et al ( 2020 ) compared different word embedding learning methods for finding software-relevant tweets. Following their guidelines, we used the same hyper-parameter settings for each word embedding learning model (i.e., Word2vec , Doc2Vec , and FastText ).…”
Section: Empirical Study Definition and Designmentioning
confidence: 99%
See 2 more Smart Citations