Large language models for software engineering: A systematic literature review

X Hou, Y Zhao, Y Liu, Z Yang, K Wang, L Li… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have significantly impacted numerous domains, notably
including Software Engineering (SE). Nevertheless, a well-rounded understanding of the …

Software testing with large language models: Survey, landscape, and vision

J Wang, Y Huang, C Chen, Z Liu… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Pre-trained large language models (LLMs) have recently emerged as a breakthrough
technology in natural language processing and artificial intelligence, with the ability to …

Expectation vs. experience: Evaluating the usability of code generation tools powered by large language models

P Vaithilingam, T Zhang, EL Glassman - Chi conference on human …, 2022 - dl.acm.org
Recent advances in Large Language Models (LLM) have made automatic code generation
possible for real-world programming tasks in general-purpose programming languages …

Codet5: Identifier-aware unified pre-trained encoder-decoder models for code understanding and generation

Y Wang, W Wang, S Joty, SCH Hoi - arXiv preprint arXiv:2109.00859, 2021 - arxiv.org
Pre-trained models for Natural Languages (NL) like BERT and GPT have been recently
shown to transfer well to Programming Languages (PL) and largely benefit a broad set of …

Program synthesis with large language models

J Austin, A Odena, M Nye, M Bosma… - arXiv preprint arXiv …, 2021 - arxiv.org
This paper explores the limits of the current generation of large language models for
program synthesis in general purpose programming languages. We evaluate a collection of …

Large language models meet nl2code: A survey

D Zan, B Chen, F Zhang, D Lu, B Wu, B Guan… - arXiv preprint arXiv …, 2022 - arxiv.org
The task of generating code from a natural language description, or NL2Code, is considered
a pressing and significant challenge in code intelligence. Thanks to the rapid development …

An empirical evaluation of using large language models for automated unit test generation

M Schäfer, S Nadi, A Eghbali… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Unit tests play a key role in ensuring the correctness of software. However, manually
creating unit tests is a laborious task, motivating the need for automation. Large Language …

No more fine-tuning? an experimental evaluation of prompt tuning in code intelligence

C Wang, Y Yang, C Gao, Y Peng, H Zhang… - Proceedings of the 30th …, 2022 - dl.acm.org
Pre-trained models have been shown effective in many code intelligence tasks. These
models are pre-trained on large-scale unlabeled corpus and then fine-tuned in downstream …

Retrieval-based prompt selection for code-related few-shot learning

N Nashid, M Sintaha, A Mesbah - 2023 IEEE/ACM 45th …, 2023 - ieeexplore.ieee.org
Large language models trained on massive code corpora can generalize to new tasks
without the need for task-specific fine-tuning. In few-shot learning, these models take as …

Natgen: generative pre-training by “naturalizing” source code

S Chakraborty, T Ahmed, Y Ding, PT Devanbu… - Proceedings of the 30th …, 2022 - dl.acm.org
Pre-trained Generative Language models (eg, PLBART, CodeT5, SPT-Code) for source
code yielded strong results on several tasks in the past few years, including code generation …