Proceedings of the 37th IEEE/ACM International Conference on Automated Software Engineering 2022
DOI: 10.1145/3551349.3559555
|View full text |Cite
|
Sign up to set email alerts
|

Few-shot training LLMs for project-specific code-summarization

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
13
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
4
1

Relationship

0
9

Authors

Journals

citations
Cited by 53 publications
(14 citation statements)
references
References 11 publications
1
13
0
Order By: Relevance
“…That said, in the mutant-test predictions, both precision and recall drop significantly for both approaches; this suggests that training data containing project-specific vocabulary and methods contribute substantially to the same project performance. This is consistent with other results showing that projects have distinct vocabulary and style, making cross project prediction difficult for many tasks [3,13]. Precision continues to be quite a bit higher than recall in the cross project setting, for both models.…”
Section: Rq1: Same Project Performancesupporting
confidence: 91%
See 1 more Smart Citation
“…That said, in the mutant-test predictions, both precision and recall drop significantly for both approaches; this suggests that training data containing project-specific vocabulary and methods contribute substantially to the same project performance. This is consistent with other results showing that projects have distinct vocabulary and style, making cross project prediction difficult for many tasks [3,13]. Precision continues to be quite a bit higher than recall in the cross project setting, for both models.…”
Section: Rq1: Same Project Performancesupporting
confidence: 91%
“…However, cloud providers increasingly provide GPU access; recently, GitHub actions announced plans to do the same for CI. 3 Indeed, GPUs are becoming more broadly accessible, including via idle GPU time or services like Google Colab. Future testing approaches or any ML for SE applications are thus increasingly realistic to deploy in practice.…”
Section: Limitations and Threatsmentioning
confidence: 99%
“…Several studies have leveraged such models for various SE tasks such as code generation [26,37,45], code repair [55,58,74], code summarization [2,36,48]. In addition, such models (specifically, ChatGPT) combine conversational capabilities with code-related tasks, allowing programmers to interact with the model.…”
Section: Related Workmentioning
confidence: 99%
“…Github Copilot uses GPT-3 for automated code generation from natural language inputs [8]. Several researchers have addressed code generation [8], [36], docstring generation [8], [60], and code repair [61], [62] problems. Bareiß et al [63] show how fewshot learning can be effective at (i) code mutation; (ii) test oracle generation from natural language documentation; and (iii) test case generation task.…”
Section: B Llms In Software Engineeringmentioning
confidence: 99%