Teach me to explain: A review of datasets for explainable natural language processing

S Wiegreffe, A Marasović - arXiv preprint arXiv:2102.12060, 2021 - arxiv.org
Explainable NLP (ExNLP) has increasingly focused on collecting human-annotated textual
explanations. These explanations are used downstream in three ways: as data …

Going beyond xai: A systematic survey for explanation-guided learning

Y Gao, S Gu, J Jiang, SR Hong, D Yu, L Zhao - ACM Computing Surveys, 2024 - dl.acm.org
As the societal impact of Deep Neural Networks (DNNs) grows, the goals for advancing
DNNs become more complex and diverse, ranging from improving a conventional model …

Language models don't always say what they think: unfaithful explanations in chain-of-thought prompting

M Turpin, J Michael, E Perez… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Large Language Models (LLMs) can achieve strong performance on many tasks by
producing step-by-step reasoning before giving a final output, often referred to as chain-of …

Distilling step-by-step! outperforming larger language models with less training data and smaller model sizes

CY Hsieh, CL Li, CK Yeh, H Nakhost, Y Fujii… - arXiv preprint arXiv …, 2023 - arxiv.org
Deploying large language models (LLMs) is challenging because they are memory
inefficient and compute-intensive for practical applications. In reaction, researchers train …

Chain-of-thought prompting elicits reasoning in large language models

J Wei, X Wang, D Schuurmans… - Advances in neural …, 2022 - proceedings.neurips.cc
We explore how generating a chain of thought---a series of intermediate reasoning steps---
significantly improves the ability of large language models to perform complex reasoning. In …

The unreliability of explanations in few-shot prompting for textual reasoning

X Ye, G Durrett - Advances in neural information processing …, 2022 - proceedings.neurips.cc
Does prompting a large language model (LLM) like GPT-3 with explanations improve in-
context learning? We study this question on two NLP tasks that involve reasoning over text …

Self-evaluation guided beam search for reasoning

Y Xie, K Kawaguchi, Y Zhao, JX Zhao… - Advances in …, 2024 - proceedings.neurips.cc
Breaking down a problem into intermediate steps has demonstrated impressive
performance in Large Language Model (LLM) reasoning. However, the growth of the …

Training language models with language feedback at scale

J Scheurer, JA Campos, T Korbak, JS Chan… - arXiv preprint arXiv …, 2023 - arxiv.org
Pretrained language models often generate outputs that are not in line with human
preferences, such as harmful text or factually incorrect summaries. Recent work approaches …

Reframing human-AI collaboration for generating free-text explanations

S Wiegreffe, J Hessel, S Swayamdipta, M Riedl… - arXiv preprint arXiv …, 2021 - arxiv.org
Large language models are increasingly capable of generating fluent-appearing text with
relatively little task-specific supervision. But can these models accurately explain …

Explanations from large language models make small reasoners better

S Li, J Chen, Y Shen, Z Chen, X Zhang, Z Li… - arXiv preprint arXiv …, 2022 - arxiv.org
Integrating free-text explanations to in-context learning of large language models (LLM) is
shown to elicit strong reasoning capabilities along with reasonable explanations. In this …