Octopack: Instruction tuning code large language models

N Muennighoff, Q Liu, A Zebaze, Q Zheng… - arXiv preprint arXiv …, 2023 - arxiv.org
Finetuning large language models (LLMs) on instructions leads to vast performance
improvements on natural language tasks. We apply instruction tuning using code …

Large language models meet nl2code: A survey

D Zan, B Chen, F Zhang, D Lu, B Wu, B Guan… - arXiv preprint arXiv …, 2022 - arxiv.org
The task of generating code from a natural language description, or NL2Code, is considered
a pressing and significant challenge in code intelligence. Thanks to the rapid development …

Codereval: A benchmark of pragmatic code generation with generative pre-trained models

H Yu, B Shen, D Ran, J Zhang, Q Zhang, Y Ma… - Proceedings of the 46th …, 2024 - dl.acm.org
Code generation models based on the pre-training and fine-tuning paradigm have been
increasingly attempted by both academia and industry, resulting in well-known industrial …

Deep Learning for Code Intelligence: Survey, Benchmark and Toolkit

Y Wan, Z Bi, Y He, J Zhang, H Zhang, Y Sui… - ACM Computing …, 2024 - dl.acm.org
Code intelligence leverages machine learning techniques to extract knowledge from
extensive code corpora, with the aim of developing intelligent tools to improve the quality …

Pangu-coder2: Boosting large language models for code with ranking feedback

B Shen, J Zhang, T Chen, D Zan, B Geng, A Fu… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models for Code (Code LLM) are flourishing. New and powerful models
are released on a weekly basis, demonstrating remarkable performance on the code …

Classeval: A manually-crafted benchmark for evaluating llms on class-level code generation

X Du, M Liu, K Wang, H Wang, J Liu, Y Chen… - arXiv preprint arXiv …, 2023 - arxiv.org
In this work, we make the first attempt to evaluate LLMs in a more challenging code
generation scenario, ie class-level code generation. We first manually construct the first …

Pangu-{\Sigma}: Towards trillion parameter language model with sparse heterogeneous computing

X Ren, P Zhou, X Meng, X Huang, Y Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
The scaling of large language models has greatly improved natural language
understanding, generation, and reasoning. In this work, we develop a system that trained a …

A survey of large language models for code: Evolution, benchmarking, and future trends

Z Zheng, K Ning, Y Wang, J Zhang, D Zheng… - arXiv preprint arXiv …, 2023 - arxiv.org
General large language models (LLMs), represented by ChatGPT, have demonstrated
significant potential in tasks such as code generation in software engineering. This has led …

Deep learning based code generation methods: A literature review

Z Yang, S Chen, C Gao, Z Li, G Li, R Lv - arXiv preprint arXiv:2303.01056, 2023 - arxiv.org
Code Generation aims at generating relevant code fragments according to given natural
language descriptions. In the process of software development, there exist a large number of …

Opencodeinterpreter: Integrating code generation with execution and refinement

T Zheng, G Zhang, T Shen, X Liu, BY Lin, J Fu… - arXiv preprint arXiv …, 2024 - arxiv.org
The introduction of large language models has significantly advanced code generation.
However, open-source models often lack the execution capabilities and iterative refinement …