Large language models for software engineering: A systematic literature review

X Hou, Y Zhao, Y Liu, Z Yang, K Wang, L Li… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have significantly impacted numerous domains, notably
including Software Engineering (SE). Nevertheless, a well-rounded understanding of the …

Lever: Learning to verify language-to-code generation with execution

A Ni, S Iyer, D Radev, V Stoyanov… - International …, 2023 - proceedings.mlr.press
The advent of large language models trained on code (code LLMs) has led to significant
progress in language-to-code generation. State-of-the-art approaches in this area combine …

Self-collaboration code generation via chatgpt

Y Dong, X Jiang, Z Jin, G Li - arXiv preprint arXiv:2304.07590, 2023 - arxiv.org
Although Large Language Models (LLMs) have demonstrated remarkable code-generation
ability, they still struggle with complex tasks. In real-world software development, humans …

Octopack: Instruction tuning code large language models

N Muennighoff, Q Liu, A Zebaze, Q Zheng… - arXiv preprint arXiv …, 2023 - arxiv.org
Finetuning large language models (LLMs) on instructions leads to vast performance
improvements on natural language tasks. We apply instruction tuning using code …

Self-planning Code Generation with Large Language Models

X Jiang, Y Dong, L Wang, F Zheng, Q Shang… - ACM Transactions on …, 2023 - dl.acm.org
Although large language models (LLMs) have demonstrated impressive ability in code
generation, they are still struggling to address the complicated intent provided by humans. It …

Self-evaluation guided beam search for reasoning

Y Xie, K Kawaguchi, Y Zhao, JX Zhao… - Advances in …, 2024 - proceedings.neurips.cc
Breaking down a problem into intermediate steps has demonstrated impressive
performance in Large Language Model (LLM) reasoning. However, the growth of the …

Self-edit: Fault-aware code editor for code generation

K Zhang, Z Li, J Li, G Li, Z Jin - arXiv preprint arXiv:2305.04087, 2023 - arxiv.org
Large language models (LLMs) have demonstrated an impressive ability to generate codes
on competitive programming tasks. However, with limited sample numbers, LLMs still suffer …

Demystifying gpt self-repair for code generation

TX Olausson, JP Inala, C Wang, J Gao… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have shown remarkable aptitude in code generation but
still struggle on challenging programming tasks. Self-repair--in which the model debugs and …

Hypothesis search: Inductive reasoning with language models

R Wang, E Zelikman, G Poesia, Y Pu, N Haber… - arXiv preprint arXiv …, 2023 - arxiv.org
Inductive reasoning is a core problem-solving capacity: humans can identify underlying
principles from a few examples, which can then be robustly generalized to novel scenarios …

Parsel🐍: Algorithmic Reasoning with Language Models by Composing Decompositions

E Zelikman, Q Huang, G Poesia… - Advances in …, 2023 - proceedings.neurips.cc
Despite recent success in large language model (LLM) reasoning, LLMs struggle with
hierarchical multi-step reasoning tasks like generating complex programs. For these tasks …