Multi-modal Large Language Models (MLLMs) have recently emerged as a significant focus in academia and industry. Despite their proficiency in general multi-modal scenarios, the …
Mathematical reasoning presents a significant challenge for Large Language Models (LLMs) due to the extensive and precise chain of reasoning required for accuracy. Ensuring …
Recent studies have shown that large language models'(LLMs) mathematical problem- solving capabilities can be enhanced by integrating external tools, such as code …
In this report, we present a series of math-specific large language models: Qwen2. 5-Math and Qwen2. 5-Math-Instruct-1.5 B/7B/72B. The core innovation of the Qwen2. 5 series lies in …
In the age of large-scale language models, benchmarks like the Massive Multitask Language Understanding (MMLU) have been pivotal in pushing the boundaries of what AI …
Continual pre-training (CPT) has been an important approach for adapting language models to specific domains or tasks. To make the CPT approach more traceable, this paper presents …
Evaluating large language models (LLMs) is challenging. Traditional ground-truth-based benchmarks fail to capture the comprehensiveness and nuance of real-world queries, while …
Y Zhang, X Chen, B Jin, S Wang, S Ji, W Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
In many scientific fields, large language models (LLMs) have revolutionized the way with which text and other modalities of data (eg, molecules and proteins) are dealt, achieving …
Despite significant advancements in the general capability of large language models (LLMs), they continue to struggle with consistent and accurate reasoning, especially in …