2019
DOI: 10.48550/arxiv.1905.03813
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

When Deep Learning Met Code Search

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
12
0

Year Published

2019
2019
2021
2021

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(12 citation statements)
references
References 0 publications
0
12
0
Order By: Relevance
“…We use FastText [27], a widely adopted [16,28,22,12] word embedding technique in software engineering. We employ FastText over the contents file (generated in Step 5) to construct a skip-gram model with the following parameters: vector size=100 (as recommended in Fasttext tutorial 10 ), epoch = 20 (higher than the default of five epochs in Fasttext tutorial to ensure possibly more effectiveness), minimum size = 2, maximum size = 5 (empirically improved result, with one unit less than the tutorial, which mentioned that other languages could have other values), and finally, left the other parameters with default values. We adopt the skip-gram model over the CBOW model because it has been observed to be more efficient with subword information [27].…”
Section: ) Generate Contents' Filesmentioning
confidence: 99%
See 2 more Smart Citations
“…We use FastText [27], a widely adopted [16,28,22,12] word embedding technique in software engineering. We employ FastText over the contents file (generated in Step 5) to construct a skip-gram model with the following parameters: vector size=100 (as recommended in Fasttext tutorial 10 ), epoch = 20 (higher than the default of five epochs in Fasttext tutorial to ensure possibly more effectiveness), minimum size = 2, maximum size = 5 (empirically improved result, with one unit less than the tutorial, which mentioned that other languages could have other values), and finally, left the other parameters with default values. We adopt the skip-gram model over the CBOW model because it has been observed to be more efficient with subword information [27].…”
Section: ) Generate Contents' Filesmentioning
confidence: 99%
“…A number of studies on Information Retrieval leverages the crowd knowledge to help developers in software development [46,47,33,22,10,13,48,15,11]. Some works [47,33] employ traditional IR techniques such as TF-IDF to recommend relevant discussions from Stack Overflow for a given context.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…In addition to the code2vec and VarMisuse tasks that we address in this paper, we believe that adversarial examples can be applied to neural code search [15,30,44] -a developer can attract users to a specific library or an open-source project by introducing code that will be disproportionately highly ranked by a neural code search model. Defending against Adversarial Examples Pruthi et al [40] proposed an approach that is related to our defense.…”
Section: Adversarial Examples In Programsmentioning
confidence: 99%
“…Neural models of code have achieved state-of-the-art performance on various tasks such as prediction of variable names and types [1,5,11,42], code summarization [2,3,20], code generation [4,13,35], code search [15,30,44], and bug finding [39,43,46].…”
Section: Introductionmentioning
confidence: 99%