2018 IEEE 25th International Conference on Software Analysis, Evolution and Reengineering (SANER) 2018
DOI: 10.1109/saner.2018.8330220
|View full text |Cite
|
Sign up to set email alerts
|

A deep neural network language model with contexts for source code

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
42
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 35 publications
(42 citation statements)
references
References 43 publications
0
42
0
Order By: Relevance
“…Second, for n-gram baselines, because the next sequence is suggested by predicting next token one at a time, the accuracy of next sequence suggestion is affected by the confounding effect of the accuracy of a single nexttoken suggestion. The highest top-1 accuracy of an n-gram LM for next code token suggestion is about 0.5 [17]. Therefore, for predicting a next code sequence containing 6 tokens (on average), the maximum top-1 accuracy is 0.5 6 ≈ 1.6%.…”
Section: Resultsmentioning
confidence: 97%
See 2 more Smart Citations
“…Second, for n-gram baselines, because the next sequence is suggested by predicting next token one at a time, the accuracy of next sequence suggestion is affected by the confounding effect of the accuracy of a single nexttoken suggestion. The highest top-1 accuracy of an n-gram LM for next code token suggestion is about 0.5 [17]. Therefore, for predicting a next code sequence containing 6 tokens (on average), the maximum top-1 accuracy is 0.5 6 ≈ 1.6%.…”
Section: Resultsmentioning
confidence: 97%
“…4). These tokens are used to initiate a set of code sequences (lines 9-12) or concatenated with the current concretized code sequences to create the new ones (lines [14][15][16][17]. The process recursively continues until the end of the template.…”
Section: Concretizing Statement Templates and Ranking Code Candimentioning
confidence: 99%
See 1 more Smart Citation
“…A major limitation of their works is that they consider source code as simple tokens of text and ignores the contextual, syntaxtual and structural dependencies. The most similar work to ours is DNN [34], however it varies in several important ways. They apply deep neural networks for source code modeling with a fixed size of context, which can only suggest the next code token, whereas our work can generate whole sequence of source code and consider variable size context.…”
Section: Related Workmentioning
confidence: 99%
“…For the training and testing of the proposed method, this work used the dataset anticipated in [5,6]. The dataset comprises of ten java projects (ant, cassandra, db40, jgit, poi, batik, antlr, itext, jts, maven).…”
Section: Datasetmentioning
confidence: 99%