This survey article proposes a clearer view of Natural Language Reasoning (NLR) in the field of Natural Language Processing (NLP), both conceptually and practically …
This survey reviews works in which language models (LMs) are augmented with reasoning skills and the ability to use tools. The former is defined as decomposing a potentially …
Reasoning is a fundamental aspect of human intelligence that plays a crucial role in activities such as problem solving, decision making, and critical thinking. In recent years …
Large language models (LLMs) have shown remarkable reasoning capabilities, especially when prompted to generate intermediate reasoning steps (eg, Chain-of-Thought, CoT) …
Reasoning, as an essential ability for complex problem-solving, can provide back-end support for various real-world applications, such as medical diagnosis, negotiation, etc. This …
Theory of Mind (ToM) $\unicode {x2014} $ the ability to reason about the mental states of other people $\unicode {x2014} $ is a key element of our social intelligence. Yet, despite …
Instruction fine-tuning has recently emerged as a promising approach for improving the zero- shot capabilities of Large Language Models (LLMs) on new tasks. This technique has …
Z Yu, L He, Z Wu, X Dai, J Chen - arXiv preprint arXiv:2310.04959, 2023 - arxiv.org
Chain-of-Thought (CoT), a step-wise and coherent reasoning chain, shows its impressive strength when used as a prompting strategy for large language models (LLM). Recent years …
In this paper, we conduct a thorough investigation into the reasoning capabilities of Large Language Models (LLMs), focusing specifically on the Open Pretrained Transformers (OPT) …