Large language models for software engineering: A systematic literature review

X Hou, Y Zhao, Y Liu, Z Yang, K Wang, L Li… - ACM Transactions on …, 2024 - dl.acm.org
Large Language Models (LLMs) have significantly impacted numerous domains, including
Software Engineering (SE). Many recent publications have explored LLMs applied to …

Igniting Language Intelligence: The Hitchhiker's Guide From Chain-of-Thought Reasoning to Language Agents

Z Zhang, Y Yao, A Zhang, X Tang, X Ma, Z He… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have dramatically enhanced the field of language
intelligence, as demonstrably evidenced by their formidable empirical performance across a …

LLMs can't plan, but can help planning in LLM-modulo frameworks

S Kambhampati, K Valmeekam, L Guan… - arXiv preprint arXiv …, 2024 - arxiv.org
There is considerable confusion about the role of Large Language Models (LLMs) in
planning and reasoning tasks. On one side are over-optimistic claims that LLMs can indeed …

Debugbench: Evaluating debugging capability of large language models

R Tian, Y Ye, Y Qin, X Cong, Y Lin, Y Pan, Y Wu… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) have demonstrated exceptional coding capability.
However, as another critical component of programming proficiency, the debugging …

Minference 1.0: Accelerating pre-filling for long-context llms via dynamic sparse attention

H Jiang, Y Li, C Zhang, Q Wu, X Luo, S Ahn… - arXiv preprint arXiv …, 2024 - arxiv.org
The computational challenges of Large Language Model (LLM) inference remain a
significant barrier to their widespread deployment, especially as prompt lengths continue to …

Data engineering for scaling language models to 128k context

Y Fu, R Panda, X Niu, X Yue, H Hajishirzi, Y Kim… - arXiv preprint arXiv …, 2024 - arxiv.org
We study the continual pretraining recipe for scaling language models' context lengths to
128K, with a focus on data engineering. We hypothesize that long context modeling, in …

Rlcoder: Reinforcement learning for repository-level code completion

Y Wang, Y Wang, D Guo, J Chen, R Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
Repository-level code completion aims to generate code for unfinished code snippets within
the context of a specified repository. Existing approaches mainly rely on retrieval-augmented …

What's Wrong with Your Code Generated by Large Language Models? An Extensive Study

S Dou, H Jia, S Wu, H Zheng, W Zhou, M Wu… - arXiv preprint arXiv …, 2024 - arxiv.org
The increasing development of large language models (LLMs) in code generation has
drawn significant attention among researchers. To enhance LLM-based code generation …

Llm-assisted code cleaning for training accurate code generators

N Jain, T Zhang, WL Chiang, JE Gonzalez… - arXiv preprint arXiv …, 2023 - arxiv.org
Natural language to code generation is an important application area of LLMs and has
received wide attention from the community. The majority of relevant studies have …

When to stop? towards efficient code generation in llms with excess token prevention

L Guo, Y Wang, E Shi, W Zhong, H Zhang… - Proceedings of the 33rd …, 2024 - dl.acm.org
Code generation aims to automatically generate code snippets that meet given natural
language requirements and plays an important role in software development. Although …