“…Recent research has delved into leveraging pretrained large language models (LLMs) from the natural language processing (NLP) field to automate program synthesis tasks, using vast-scale code corpus data mined from open-source repositories. Notably, there are several prominent examples of such pretrained models including the encoder-only CodeBERT (Feng et al, 2020), decoder-only CodeGPT (Lu et al, 2021), Code-Gen (Nijkamp et al, 2022), PaLM-Coder (Chowdhery et al, 2022), PanGu-Coder (Christopoulou et al, 2022), CodeGeex (Zheng et al, 2023), andSantaCoder (Allal et al, 2023), as well as encoder-decoder transformer architectures like PLABRT (Ahmad et al, 2021) and CodeT5 (Wang et al, 2021). These pretrained probabilistic language (PL) models are already capable of generating code that appears visually impressive and well-structured.…”