2020
DOI: 10.48550/arxiv.2005.08025
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

IntelliCode Compose: Code Generation Using Transformer

Abstract: In software development through integrated development environments (IDEs), code completion is one of the most widely used features. Nevertheless, majority of integrated development environments only support completion of methods and APIs, or arguments. In this paper, we introduce IntelliCode Compose -a general-purpose multilingual code completion tool which is capable of predicting sequences of code tokens of arbitrary types, generating up to entire lines of syntactically correct code. It leverages state-of-t… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
33
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
7

Relationship

1
6

Authors

Journals

citations
Cited by 16 publications
(33 citation statements)
references
References 19 publications
0
33
0
Order By: Relevance
“…Pre-Trained Models for Programming Languages Inspired by the big success of pre-training in NLP (Devlin et al, 2018;Yang et al, 2019;Raffel et al, 2019), pre-trained models for programming languages also promotes the development of code intelligence (Kanade et al, 2019;Feng et al, 2020;Karampatsis & Sutton, 2020;Svyatkovskiy et al, 2020;Buratti et al, 2020). Kanade et al (2019) pre-train a BERT model on a massive corpus of Python source codes by masked language modeling and next sentence prediction objectives.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Pre-Trained Models for Programming Languages Inspired by the big success of pre-training in NLP (Devlin et al, 2018;Yang et al, 2019;Raffel et al, 2019), pre-trained models for programming languages also promotes the development of code intelligence (Kanade et al, 2019;Feng et al, 2020;Karampatsis & Sutton, 2020;Svyatkovskiy et al, 2020;Buratti et al, 2020). Kanade et al (2019) pre-train a BERT model on a massive corpus of Python source codes by masked language modeling and next sentence prediction objectives.…”
Section: Related Workmentioning
confidence: 99%
“…The success of pre-trained models in NLP also promotes the development of pre-trained models for programming language. Existing works (Kanade et al, 2019;Karampatsis & Sutton, 2020;Feng et al, 2020;Svyatkovskiy et al, 2020;Buratti et al, 2020) regard a source code as a sequence of tokens and pre-train models on source code to support code-related tasks such as code search, code completion, code summarization, etc. However, previous works only utilize source code for pre-training, while ignoring the inherent structure of code.…”
Section: Introductionmentioning
confidence: 99%
“…In recent years, several efforts have focused on the use of AI and machine learning techniques for various tasks related to software engineering, including code completion [15,41,75,84], code classification [49,68], API recommendation [16,33], variable and method naming [3,5], type inference [39,93], bug detection and repair [25,40,71,74,89,95], comment description and generation [4,44,48,65,80,91], code change summarization [66], and code clone detection [96]. A significant portion of this work is recounted in Allamanis et al 's survey of the area [2].…”
Section: Ai Techniques For Software Engineeringmentioning
confidence: 99%
“…Svyatkovskiy et al [15] introduced IntelliCode Compose, a general-purpose multilingual code completion tool capable of predicting code sequences of arbitrary token types. They do not leverage high-level structural representation, such as AST, and use subtokens to overcome the out of vocabulary problem.…”
Section: Related Workmentioning
confidence: 99%
“…Although the performance of code completion techniques substantially improved over time, the type of support they provide to developers has not evolved at the same pace, and are mostly only capable of predicting a single token. Only a few recent studies focus on predicting multiple contiguous tokens [14], [15].…”
Section: Introductionmentioning
confidence: 99%