Distilling LLMs' Decomposition Abilities into Compact Language Models

D Tarasov, K Shridhar - arXiv preprint arXiv:2402.01812, 2024 - arxiv.org
Large Language Models (LLMs) have demonstrated proficiency in their reasoning abilities,
yet their large size presents scalability challenges and limits any further customization. In …

Small language models fine-tuned to coordinate larger language models improve complex reasoning

G Juneja, S Dutta, S Chakrabarti, S Manchanda… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) prompted to generate chain-of-thought (CoT) exhibit
impressive reasoning capabilities. Recent attempts at prompt decomposition toward solving …

First Step Advantage: Importance of Starting Right in Multi-Step Reasoning

K Jain, K Shridhar - arXiv preprint arXiv:2311.07945, 2023 - arxiv.org
Large Language Models (LLMs) can solve complex reasoning tasks by generating
rationales for their predictions. Distilling these capabilities into a smaller, compact model can …

Divide-or-Conquer? Which Part Should You Distill Your LLM?

Z Wu, H Bai, A Zhang, J Gu, VG Vydiswaran… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent methods have demonstrated that Large Language Models (LLMs) can solve
reasoning tasks better when they are encouraged to solve subtasks of the main task first. In …

Mixture-of-Skills: Learning to Optimize Data Usage for Fine-Tuning Large Language Models

M Wu, TT Vu, L Qu, G Haffari - arXiv preprint arXiv:2406.08811, 2024 - arxiv.org
Large language models (LLMs) are typically fine-tuned on diverse and extensive datasets
sourced from various origins to develop a comprehensive range of skills, such as writing …

Beyond Accuracy: Evaluating the Reasoning Behavior of Large Language Models--A Survey

P Mondorf, B Plank - arXiv preprint arXiv:2404.01869, 2024 - arxiv.org
Large language models (LLMs) have recently shown impressive performance on tasks
involving reasoning, leading to a lively debate on whether these models possess reasoning …

Adaptive-solver framework for dynamic strategy selection in large language model reasoning

J Zhou, W Zhong, Y Wang, J Wang - arXiv preprint arXiv:2310.01446, 2023 - arxiv.org
Large Language Models (LLMs) are showcasing impressive ability in handling complex
reasoning tasks. In real-world situations, problems often span a spectrum of complexities …

Learning to Reduce: Towards Improving Performance of Large Language Models on Structured Data

Y Lee, S Kim, RA Rossi, T Yu, X Chen - arXiv preprint arXiv:2407.02750, 2024 - arxiv.org
Large Language Models (LLMs) have been achieving competent performance on a wide
range of downstream tasks, yet existing work shows that inference on structured data is …

AMSP: Super-Scaling LLM Training via Advanced Model States Partitioning

Q Chen, Q Hu, Z Ye, G Wang, P Sun, Y Wen… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) have demonstrated impressive performance across various
downstream tasks. When training these models, there is a growing inclination to process …

Orca 2: Teaching small language models how to reason

A Mitra, L Del Corro, S Mahajan, A Codas… - arXiv preprint arXiv …, 2023 - arxiv.org
Orca 1 learns from rich signals, such as explanation traces, allowing it to outperform
conventional instruction-tuned models on benchmarks like BigBench Hard and AGIEval. In …