Challenges and applications of large language models

J Kaddour, J Harris, M Mozes, H Bradley… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) went from non-existent to ubiquitous in the machine
learning discourse within a few years. Due to the fast pace of the field, it is difficult to identify …

Natural language reasoning, a survey

F Yu, H Zhang, P Tiwari, B Wang - ACM Computing Surveys, 2024 - dl.acm.org
This survey article proposes a clearer view of Natural Language Reasoning (NLR) in the
field of Natural Language Processing (NLP), both conceptually and practically …

Augmented language models: a survey

G Mialon, R Dessì, M Lomeli, C Nalmpantis… - arXiv preprint arXiv …, 2023 - arxiv.org
This survey reviews works in which language models (LMs) are augmented with reasoning
skills and the ability to use tools. The former is defined as decomposing a potentially …

Towards reasoning in large language models: A survey

J Huang, KCC Chang - arXiv preprint arXiv:2212.10403, 2022 - arxiv.org
Reasoning is a fundamental aspect of human intelligence that plays a crucial role in
activities such as problem solving, decision making, and critical thinking. In recent years …

Reasoning with language model is planning with world model

S Hao, Y Gu, H Ma, JJ Hong, Z Wang, DZ Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have shown remarkable reasoning capabilities, especially
when prompted to generate intermediate reasoning steps (eg, Chain-of-Thought, CoT) …

Reasoning with language model prompting: A survey

S Qiao, Y Ou, N Zhang, X Chen, Y Yao, S Deng… - arXiv preprint arXiv …, 2022 - arxiv.org
Reasoning, as an essential ability for complex problem-solving, can provide back-end
support for various real-world applications, such as medical diagnosis, negotiation, etc. This …

Minding language models'(lack of) theory of mind: A plug-and-play multi-character belief tracker

M Sclar, S Kumar, P West, A Suhr, Y Choi… - arXiv preprint arXiv …, 2023 - arxiv.org
Theory of Mind (ToM) $\unicode {x2014} $ the ability to reason about the mental states of
other people $\unicode {x2014} $ is a key element of our social intelligence. Yet, despite …

Evaluating the zero-shot robustness of instruction-tuned language models

J Sun, C Shaib, BC Wallace - arXiv preprint arXiv:2306.11270, 2023 - arxiv.org
Instruction fine-tuning has recently emerged as a promising approach for improving the zero-
shot capabilities of Large Language Models (LLMs) on new tasks. This technique has …

Towards better chain-of-thought prompting strategies: A survey

Z Yu, L He, Z Wu, X Dai, J Chen - arXiv preprint arXiv:2310.04959, 2023 - arxiv.org
Chain-of-Thought (CoT), a step-wise and coherent reasoning chain, shows its impressive
strength when used as a prompting strategy for large language models (LLM). Recent years …

Opt-r: Exploring the role of explanations in finetuning and prompting for reasoning skills of large language models

B AlKhamissi, S Verma, P Yu, Z Jin… - arXiv preprint arXiv …, 2023 - arxiv.org
In this paper, we conduct a thorough investigation into the reasoning capabilities of Large
Language Models (LLMs), focusing specifically on the Open Pretrained Transformers (OPT) …