2021
DOI: 10.48550/arxiv.2103.07115
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

An Empirical Study on the Usage of BERT Models for Code Completion

Abstract: Code completion is one of the main features of modern Integrated Development Environments (IDEs). Its objective is to speed up code writing by predicting the next code token(s) the developer is likely to write. Research in this area has substantially bolstered the predictive performance of these techniques. However, the support to developers is still limited to the prediction of the next few tokens to type. In this work, we take a step further in this direction by presenting a large-scale empirical study aimed… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
4
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
3
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 29 publications
(58 reference statements)
0
4
0
Order By: Relevance
“…Similar to traditional assistants, intelligent ones allow developers to explore APIs by displaying a list of all available methods and attributes. However, these results are typically ranked by relevance rather than in alphabetical order [32]. With the help of these data, they can determine the developer's intention and generate a proposal that is as relevant as possible to the developer.…”
Section: Functionalities Sourcesmentioning
confidence: 99%
“…Similar to traditional assistants, intelligent ones allow developers to explore APIs by displaying a list of all available methods and attributes. However, these results are typically ranked by relevance rather than in alphabetical order [32]. With the help of these data, they can determine the developer's intention and generate a proposal that is as relevant as possible to the developer.…”
Section: Functionalities Sourcesmentioning
confidence: 99%
“…3) Empirical Studies on Auto-Completion Models: Ciniselli et al [29,30] analyzed the performance of two language models for text namely, T5 [17] and RoBERTa [16], for completing code in three granularity levels; single-token, line, and block. The authors included two datasets, containing Java methods and Android app methods from open-source GitHub repositories.…”
Section: Background and Related Workmentioning
confidence: 99%
“…Both approaches have their flaws. Due to their nature, DSL languages match specific domains, and will never become general-purpose tools; generative LLMs have difficulty extracting complex coding patterns from code corpora, and often generate codes riddled with syntax or semantic errors [12], [13]. The results returned by either model are seldom predictable or replicable.…”
Section: B Codex Copilot Gpt-3mentioning
confidence: 99%