Qwen2. 5-coder technical report

B Hui, J Yang, Z Cui, J Yang, D Liu, L Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
In this report, we introduce the Qwen2. 5-Coder series, a significant upgrade from its
predecessor, CodeQwen1. 5. This series includes six models: Qwen2. 5-Coder-(0.5 B/1.5 …

Mdeval: Massively multilingual code debugging

S Liu, L Chai, J Yang, J Shi, H Zhu, L Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Code large language models (LLMs) have made significant progress in code debugging by
directly generating the correct code based on the buggy code snippet. Programming …

Evaluating and aligning codellms on human preference

J Yang, J Yang, K Jin, Y Miao, L Zhang, L Yang… - arXiv preprint arXiv …, 2024 - arxiv.org
Code large language models (codeLLMs) have made significant strides in code generation.
Most previous code-related benchmarks, which consist of various programming exercises …

Fullstack bench: Evaluating llms as full stack coder

S Liu, H Zhu, J Liu, S Xin, A Li, R Long, L Chen… - arXiv preprint arXiv …, 2024 - arxiv.org
As the capabilities of code large language models (LLMs) continue to expand, their
applications across diverse code intelligence domains are rapidly increasing. However …