What's Wrong with Your Code Generated by Large Language Models? An Extensive Study

S Dou, H Jia, S Wu, H Zheng, W Zhou, M Wu - arXiv preprint arXiv , 2024 - arxiv.org
The increasing development of large language models (LLMs) in code generation has
drawn significant attention among researchers. To enhance LLM-based code generation 

A deep dive into large language models for automated bug localization and repair

SB Hossain, N Jiang, Q Zhou, X Li, WH Chiang - Proceedings of the , 2024 - dl.acm.org
Large language models (LLMs) have shown impressive effectiveness in various software
engineering tasks, including automated program repair (APR). In this study, we take a deep 

Stepcoder: Improve code generation with reinforcement learning from compiler feedback

S Dou, Y Liu, H Jia, L Xiong, E Zhou, W Shen - arXiv preprint arXiv , 2024 - arxiv.org
The advancement of large language models (LLMs) has significantly propelled the field of
code generation. Previous work integrated reinforcement learning (RL) with compiler 

Towards a unified view of preference learning for large language models: A survey

B Gao, F Song, Y Miao, Z Cai, Z Yang, L Chen - arXiv preprint arXiv , 2024 - arxiv.org
Large Language Models (LLMs) exhibit remarkably powerful capabilities. One of the crucial
factors to achieve success is aligning the LLM's output with human preferences. This 

Transformers in source code generation: A comprehensive survey

H Ghaemi, Z Alizadehsani, A Shahraki - Journal of Systems , 2024 - Elsevier
Transformers have revolutionized natural language processing (NLP) and have had a huge
impact on automating tasks. Recently, transformers have led to the development of powerful 
ô5

Agents in software engineering: Survey, landscape, and vision

Y Wang, W Zhong, Y Huang, E Shi, M Yang - arXiv preprint arXiv , 2024 - arxiv.org
In recent years, Large Language Models (LLMs) have achieved remarkable success and
have been widely used in various downstream tasks, especially in the tasks of the software 

From Code to Correctness: Closing the Last Mile of Code Generation with Hierarchical Debugging

Y Shi, S Wang, C Wan, X Gu - arXiv preprint arXiv:2410.01215, 2024 - arxiv.org
While large language models have made significant strides in code generation, the pass
rate of the generated code is bottlenecked on subtle errors, often requiring human 

A survey of neural code intelligence: Paradigms, advances and beyond

Q Sun, Z Chen, F Xu, K Cheng, C Ma, Z Yin - arXiv preprint arXiv , 2024 - arxiv.org
Neural Code Intelligence--leveraging deep learning to understand, generate, and optimize
code--holds immense potential for transformative impacts on the whole society. Bridging the 

Fuzzing JavaScript Interpreters with Coverage-Guided Reinforcement Learning for LLM-Based Mutation

J Eom, S Jeong, T Kwon - Proceedings of the 33rd ACM SIGSOFT , 2024 - dl.acm.org
JavaScript interpreters, crucial for modern web browsers, require an effective fuzzing
method to identify security-related bugs. However, the strict grammatical requirements for 

A Survey on Large Language Models for Code Generation

J Jiang, F Wang, J Shen, S Kim, S Kim - arXiv preprint arXiv:2406.00515, 2024 - arxiv.org
Large Language Models (LLMs) have garnered remarkable advancements across diverse
code-related tasks, known as Code LLMs, particularly in code generation that generates