2021 IEEE/ACM 43rd International Conference on Software Engineering (ICSE) 2021
DOI: 10.1109/icse43902.2021.00026
|View full text |Cite
|
Sign up to set email alerts
|

Code Prediction by Feeding Trees to Transformers

Abstract: Code prediction, more specifically autocomplete, has become an essential feature in modern IDEs. Autocomplete is more effective when the desired next token is at (or close to) the top of the list of potential completions offered by the IDE at cursor position. This is where the strength of the underlying machine learning system that produces a ranked order of potential completions comes into play.We advance the state-of-the-art in the accuracy of code prediction (next token prediction) used in autocomplete syst… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
103
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
3
3

Relationship

0
9

Authors

Journals

citations
Cited by 127 publications
(109 citation statements)
references
References 38 publications
0
103
0
Order By: Relevance
“…In traditional grammar-based generation of text [7] or code [21,38,2,5], the CFG is followed by sequentially expanding the left-most, bottom-most non-terminal symbol, using one of the production rules in R. GRAMMFORMER changes this and instead selects which (if any) non-terminal symbol to expand. Similar to recent works [38,5,16], GRAMMFORMER loosens the CFG assumptions but retains many aspects, discussed next. Alg.…”
Section: Grammformermentioning
confidence: 52%
See 1 more Smart Citation
“…In traditional grammar-based generation of text [7] or code [21,38,2,5], the CFG is followed by sequentially expanding the left-most, bottom-most non-terminal symbol, using one of the production rules in R. GRAMMFORMER changes this and instead selects which (if any) non-terminal symbol to expand. Similar to recent works [38,5,16], GRAMMFORMER loosens the CFG assumptions but retains many aspects, discussed next. Alg.…”
Section: Grammformermentioning
confidence: 52%
“…One of the most successful applications of LMCs is code completion [33,15] and transformer language models have been recently shown exceptional performance at the task being able to predict relatively long sequences of code tokens [34]. Grammar-based code completion and generation has been researched with neural [21,38,16] and non-neural models [5], always expanding the left-most, bottom-most non-terminal. In contrast to GRAMMFORMERs, all these code completion models target the generation of complete code without the ability to create sketches.…”
Section: Related Workmentioning
confidence: 99%
“…V. RELATED WORK Abstract Syntax Trees (AST) have been used extensively in the literature [6], [7], [14]- [16], [21], [28]. In their work, Zhang et al [14] use sub-trees extracted from the AST with tree-based CNN to generate code vectors.…”
Section: Evaluation and Resultsmentioning
confidence: 99%
“…coding [Kim et al, 2021], and VR interaction [David-John et al, 2021]. These approaches, however, have posed problems since their conception because they can be frustrating or detrimental to completing tasks if the computer cannot correctly predict what the user intended [Olteanu et al, 2020;Yang et al, 2020].…”
Section: Dimension 2: Computer Assistancementioning
confidence: 99%