2023
DOI: 10.48550/arxiv.2303.05510
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Planning with Large Language Models for Code Generation

Abstract: Existing large language model-based code generation pipelines typically use beam search or sampling algorithms during the decoding process. Although the programs they generate achieve high token-matching-based scores, they often fail to compile or generate incorrect outputs. The main reason is that conventional Transformer decoding algorithms may not be the best choice for code generation. In this work, we propose a novel Transformer decoding algorithm, Planning-Guided Transformer Decoding (PG-TD), that uses a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 30 publications
0
1
0
Order By: Relevance
“…Among decoding-time approaches, many are focused on optimization, either through beam search [Meister et al, 2020] and heuristic search [Lu et al, 2021, Zhang et al, 2023b, or through gradientbased optimization in embedding space [Dathathri et al, 2019, Kumar et al, 2021. Other approaches focus on sampling from a constrained or modified distribution [Zhang et al, 2023a], including naive rejection sampling [Poesia et al, 2022], but also more sophisticated Markov Chain Monte Carlo (MCMC) samplers [Miao et al, 2019, Hie et al, 2022 that make use of specialized proposal distributions [Zhang et al, 2020] or the gradients of continuous embeddings [Qin et al, 2022, Kumar et al, 2022.…”
Section: Related Work and Discussionmentioning
confidence: 99%
“…Among decoding-time approaches, many are focused on optimization, either through beam search [Meister et al, 2020] and heuristic search [Lu et al, 2021, Zhang et al, 2023b, or through gradientbased optimization in embedding space [Dathathri et al, 2019, Kumar et al, 2021. Other approaches focus on sampling from a constrained or modified distribution [Zhang et al, 2023a], including naive rejection sampling [Poesia et al, 2022], but also more sophisticated Markov Chain Monte Carlo (MCMC) samplers [Miao et al, 2019, Hie et al, 2022 that make use of specialized proposal distributions [Zhang et al, 2020] or the gradients of continuous embeddings [Qin et al, 2022, Kumar et al, 2022.…”
Section: Related Work and Discussionmentioning
confidence: 99%