Proceedings of the 44th International Conference on Software Engineering 2022
DOI: 10.1145/3510003.3510621
|View full text |Cite
|
Sign up to set email alerts
|

Using pre-trained models to boost code review automation

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
6
4

Relationship

0
10

Authors

Journals

citations
Cited by 64 publications
(6 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…As representative of transformers [1], we adopt the T5 proposed by Raffel et al [20], that has been already used in SE to automate code-related tasks [9], [13], [14], [58], [59]. Masks X% of tokens (usually 15%) in the instance (e.g., a function) and asks the model to guess the masked tokens based on their bidirectional context.…”
Section: A Transformer Modelmentioning
confidence: 99%
“…As representative of transformers [1], we adopt the T5 proposed by Raffel et al [20], that has been already used in SE to automate code-related tasks [9], [13], [14], [58], [59]. Masks X% of tokens (usually 15%) in the instance (e.g., a function) and asks the model to guess the masked tokens based on their bidirectional context.…”
Section: A Transformer Modelmentioning
confidence: 99%
“…Thongtanunam et al [41] further introduces advanced Transformer architecture and a Byte-Pair Encoding (BPE) approach to handle the Out-Of-Vocabulary and long sequence problems. To better learn code properties, pre-training techniques are increasingly adopted in the code review scenario [11,18,24,44,54]. Hong et al [18] proposes a CodeT5-based approach to recommend code review comments automatically.…”
Section: Nmt Models For Code Generationmentioning
confidence: 99%
“…3) Fine-tuning: We fine-tune the best pre-trained model (T5 NL+DF ) with the best learning rate strategy (ST-LR) on D FT-train . We use early stopping to avoid overfitting [29], [38]: We save a checkpoint every 10k steps and compute the BLEU-4 score on the evaluation set every 100k steps. When the 100k steps do not lead to an improvement, we stop the training procedure, and we keep the last model.…”
Section: Training T5 For Generating Dockerfilesmentioning
confidence: 99%