2021 IEEE/ACM 18th International Conference on Mining Software Repositories (MSR) 2021
DOI: 10.1109/msr52588.2021.00024
|View full text |Cite
|
Sign up to set email alerts
|

An Empirical Study on the Usage of BERT Models for Code Completion

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
22
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
3
2
2

Relationship

1
6

Authors

Journals

citations
Cited by 47 publications
(23 citation statements)
references
References 36 publications
0
22
0
Order By: Relevance
“…In this work, we extend our MSR 2021 paper [22] by showing that the T5 substantially overcomes the performance of the RoBERTa model, being able to correctly predict even entire code blocks, something that we found to be not achievable with RoBERTa. As in [22], we focus on three code prediction scenarios: (i) token-level predictions, namely classic code completion in which the model is used to guess the last n tokens in a statement the developer started writing; (ii) construct-level predictions, in which the model is used to predict specific code constructs (e.g., the condition of an if statement) that can be particularly useful to developers while writing code; and (iii) block-level predictions, with the masked code spanning one or more entire statements composing a code block (e.g., the iterated block of a for loop).…”
Section: Introductionmentioning
confidence: 64%
See 3 more Smart Citations
“…In this work, we extend our MSR 2021 paper [22] by showing that the T5 substantially overcomes the performance of the RoBERTa model, being able to correctly predict even entire code blocks, something that we found to be not achievable with RoBERTa. As in [22], we focus on three code prediction scenarios: (i) token-level predictions, namely classic code completion in which the model is used to guess the last n tokens in a statement the developer started writing; (ii) construct-level predictions, in which the model is used to predict specific code constructs (e.g., the condition of an if statement) that can be particularly useful to developers while writing code; and (iii) block-level predictions, with the masked code spanning one or more entire statements composing a code block (e.g., the iterated block of a for loop).…”
Section: Introductionmentioning
confidence: 64%
“…Such research has allowed to move from simple alphabetically ranked lists of recommendations for completing what a developer is typing (e.g., a list of possible method calls matching what has been typed by the developer) to "in-telligent" completions considering the context surrounding the code [17], [66], the history of code changes [66], and/or coding patterns mined from software repositories [9], [36], [38], [59], [60], [61], [72]. Last, but not least, Deep Learning (DL) models have been applied to code completion [7], [22], [45], [47], [68], [77], setting new standards in terms of prediction performance. Although the performance of code completion techniques have substantially improved over time, the type of support they provide to developers has not evolved at the same pace.…”
Section: Introductionmentioning
confidence: 99%
See 2 more Smart Citations
“…For examples, Harer et al (2018), Ben-Nun, Jakobovits, and Hoefler (2018), and Zuo et al (2019 apply the word2vec model (Le and Mikolov 2014) to learn the embeddings of program tokens. Feng et al (2020), Wang et al (2020) and Ciniselli et al (2021) use a pre-trained BERT model to encode programs. Such sequence-based methods are easy to use and can benefit largely from the NLP community.…”
Section: Related Workmentioning
confidence: 99%