Proceedings of the 44th International Conference on Software Engineering 2022
DOI: 10.1145/3510003.3510049
|View full text |Cite
|
Sign up to set email alerts
|

Multilingual training for software engineering

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
1

Year Published

2023
2023
2023
2023

Publication Types

Select...
6
2
1

Relationship

0
9

Authors

Journals

citations
Cited by 33 publications
(26 citation statements)
references
References 32 publications
0
25
1
Order By: Relevance
“…Chen et al [86] investigated the proposal by Ahmed and Devanbu [87] to pre-train DL models on multiple programming languages. The authors reported that multilingual models have worst performance as compared to monolingual ones.…”
Section: Related Workmentioning
confidence: 99%
“…Chen et al [86] investigated the proposal by Ahmed and Devanbu [87] to pre-train DL models on multiple programming languages. The authors reported that multilingual models have worst performance as compared to monolingual ones.…”
Section: Related Workmentioning
confidence: 99%
“…Data Duplication. Prior studies [4,39] found that duplicated data across training and testing could lead to unrealistic model performance. However, the prior studies only focus on code completion and code summarization tasks.…”
Section: (Rq2) How Reliable Are Automated Code Generation Approaches?mentioning
confidence: 99%
“…Code generation models have been applied to a variety of tasks, including test generation [19], docstring generation [20], code search [17,21], type inference [22,23,24], and more [25]. We focus on the natural-language-to-code task (NL2Code): given the description of a function in natural language, complete the function body.…”
Section: The Natural Language To Code Taskmentioning
confidence: 99%
“…Other tasks. Although we focus specifically on benchmarks for the code generation task, there are many other tasks that have been used to evaluate code generation models, including generating unit tests from code [19], code search [17,21], and type inference [22,23,24]. Lu et al [20] propose a suite of evaluation datasets for ten tasks, including code translation, docstring generation, and code summarization.…”
Section: Related Workmentioning
confidence: 99%