A survey of deep learning for mathematical reasoning

P Lu, L Qiu, W Yu, S Welleck, KW Chang - arXiv preprint arXiv:2212.10535, 2022 - arxiv.org
Mathematical reasoning is a fundamental aspect of human intelligence and is applicable in
various fields, including science, engineering, finance, and everyday life. The development …

Palm: Scaling language modeling with pathways

A Chowdhery, S Narang, J Devlin, M Bosma… - Journal of Machine …, 2023 - jmlr.org
Large language models have been shown to achieve remarkable performance across a
variety of natural language tasks using few-shot learning, which drastically reduces the …

Chain-of-thought prompting elicits reasoning in large language models

J Wei, X Wang, D Schuurmans… - Advances in neural …, 2022 - proceedings.neurips.cc
We explore how generating a chain of thought---a series of intermediate reasoning steps---
significantly improves the ability of large language models to perform complex reasoning. In …

Wizardmath: Empowering mathematical reasoning for large language models via reinforced evol-instruct

H Luo, Q Sun, C Xu, P Zhao, J Lou, C Tao… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs), such as GPT-4, have shown remarkable performance in
natural language processing (NLP) tasks, including challenging mathematical reasoning …

Active prompting with chain-of-thought for large language models

S Diao, P Wang, Y Lin, T Zhang - arXiv preprint arXiv:2302.12246, 2023 - arxiv.org
The increasing scale of large language models (LLMs) brings emergent abilities to various
complex tasks requiring reasoning, such as arithmetic and commonsense reasoning. It is …

Datasets for large language models: A comprehensive survey

Y Liu, J Cao, C Liu, K Ding, L Jin - arXiv preprint arXiv:2402.18041, 2024 - arxiv.org
This paper embarks on an exploration into the Large Language Model (LLM) datasets,
which play a crucial role in the remarkable advancements of LLMs. The datasets serve as …

Automatic prompt augmentation and selection with chain-of-thought from labeled data

KS Shum, S Diao, T Zhang - arXiv preprint arXiv:2302.12822, 2023 - arxiv.org
Chain-of-thought prompting (CoT) advances the reasoning abilities of large language
models (LLMs) and achieves superior performance in arithmetic, commonsense, and …

Learning to reason deductively: Math word problem solving as complex relation extraction

Z Jie, J Li, W Lu - arXiv preprint arXiv:2203.10316, 2022 - arxiv.org
Solving math word problems requires deductive reasoning over the quantities in the text.
Various recent research efforts mostly relied on sequence-to-sequence or sequence-to-tree …

Solving math word problems via cooperative reasoning induced language models

X Zhu, J Wang, L Zhang, Y Zhang, Y Huang… - arXiv preprint arXiv …, 2022 - arxiv.org
Large-scale pre-trained language models (PLMs) bring new opportunities to challenging
problems, especially those that need high-level intelligence, such as the math word problem …

Let gpt be a math tutor: Teaching math word problem solvers with customized exercise generation

Z Liang, W Yu, T Rajpurohit, P Clark, X Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
In this paper, we present a novel approach for distilling math word problem solving
capabilities from large language models (LLMs) into smaller, more efficient student models …