2023
DOI: 10.1109/tse.2022.3183297
|View full text |Cite
|
Sign up to set email alerts
|

Using Transfer Learning for Code-Related Tasks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
11
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 24 publications
(11 citation statements)
references
References 59 publications
0
11
0
Order By: Relevance
“…As representative of transformers [1], we adopt the T5 proposed by Raffel et al [20], that has been already used in SE to automate code-related tasks [9], [13], [14], [58], [59]. Masks X% of tokens (usually 15%) in the instance (e.g., a function) and asks the model to guess the masked tokens based on their bidirectional context.…”
Section: A Transformer Modelmentioning
confidence: 99%
See 1 more Smart Citation
“…As representative of transformers [1], we adopt the T5 proposed by Raffel et al [20], that has been already used in SE to automate code-related tasks [9], [13], [14], [58], [59]. Masks X% of tokens (usually 15%) in the instance (e.g., a function) and asks the model to guess the masked tokens based on their bidirectional context.…”
Section: A Transformer Modelmentioning
confidence: 99%
“…Then, a labeled dataset mapping English sentences to their corresponding German translation can be used to fine-tune the model. Several works applying DL to SE report boost of performance 1 provided by pre-training in the automation of code-related tasks [11], [13], [14]. However, little is known about (i) the circumstances in which pre-training actually helps, and (ii) the impact of the specific pre-training objective(s) adopted on the performance of transformers when automating code-related tasks.…”
Section: Introductionmentioning
confidence: 99%
“…Mastropaolo et al [104] propose pre-trained text-to-text transfer transformer (T5) to address four code-related tasks, namely automatic bug fixing, injection of code mutants, generation of assert statements in test methods, and code summarization. They apply BFP small and BFP medium datasets to train and evaluate the bug-fixing task, and then compare other state-of-art learningbased APR tools on the same benchmark.…”
Section: Universalmentioning
confidence: 99%
“…We run our specificationinference tool on them and, after a filtering procedure where duplicates and invalid Dockerfiles are removed, we end up with a set of 670,982 unique pairs HLS, Dockerfile . We use this dataset to train and test a state-of-the-art DL model, the Text-to-Text Transfer Transformer (T5) [17], which has been proven effective when supporting several coding tasks [13], [14], following the same pipeline defined in the literature. We compare the DL-based approach with two Information Retrieval (IR)-based approaches (i.e., less complex and lessresource-requiring alternatives), and we check to what extent, given a HLS, the output Dockerfiles of the three techniques: (i) meet the input requirements, (ii) are similar to the target Dockerfile, and (iii) allow to build a Docker image similar to the target one.…”
Section: Introductionmentioning
confidence: 99%
“…The stop criterion we adopt, which is the one currently used for coding tasks [14], is based on the convergence in terms of BLEU-4 score. However, considering our results, it seems to be ineffective in the evaluated context.…”
Section: Introductionmentioning
confidence: 99%