Contrastive decoding: Open-ended text generation as optimization

XL Li, A Holtzman, D Fried, P Liang, J Eisner… - arXiv preprint arXiv …, 2022 - arxiv.org
Given a language model (LM), maximum probability is a poor decoding objective for open-
ended generation, because it produces short and repetitive text. On the other hand …

Contrastive decoding improves reasoning in large language models

S O'Brien, M Lewis - arXiv preprint arXiv:2309.09117, 2023 - arxiv.org
We demonstrate that Contrastive Decoding--a simple, computationally light, and training-
free text generation method proposed by Li et al 2022--achieves large out-of-the-box …

Neurologic a* esque decoding: Constrained text generation with lookahead heuristics

X Lu, S Welleck, P West, L Jiang, J Kasai… - arXiv preprint arXiv …, 2021 - arxiv.org
The dominant paradigm for neural text generation is left-to-right decoding from
autoregressive language models. Constrained or controllable generation under complex …

Rankgen: Improving text generation with large ranking models

K Krishna, Y Chang, J Wieting, M Iyyer - arXiv preprint arXiv:2205.09726, 2022 - arxiv.org
Given an input sequence (or prefix), modern language models often assign high
probabilities to output sequences that are repetitive, incoherent, or irrelevant to the prefix; as …

Text generation by learning from demonstrations

RY Pang, H He - arXiv preprint arXiv:2009.07839, 2020 - arxiv.org
Current approaches to text generation largely rely on autoregressive models and maximum
likelihood estimation. This paradigm leads to (i) diverse but low-quality samples due to …

A plug-and-play method for controlled text generation

D Pascual, B Egressy, C Meister, R Cotterell… - arXiv preprint arXiv …, 2021 - arxiv.org
Large pre-trained language models have repeatedly shown their ability to produce fluent
text. Yet even when starting from a prompt, generation can continue in many plausible …

Contrastive search is what you need for neural text generation

Y Su, N Collier - arXiv preprint arXiv:2210.14140, 2022 - arxiv.org
Generating text with autoregressive language models (LMs) is of great importance to many
natural language processing (NLP) applications. Previous solutions for this task often …

Follow the wisdom of the crowd: Effective text generation via minimum Bayes risk decoding

M Suzgun, L Melas-Kyriazi, D Jurafsky - arXiv preprint arXiv:2211.07634, 2022 - arxiv.org
In open-ended natural-language generation, existing text decoding methods typically
struggle to produce text which is both diverse and high-quality. Greedy and beam search are …

Factuality enhanced language models for open-ended text generation

N Lee, W Ping, P Xu, M Patwary… - Advances in …, 2022 - proceedings.neurips.cc
Pretrained language models (LMs) are susceptible to generate text with nonfactual
information. In this work, we measure and improve the factual accuracy of large-scale LMs …

DExperts: Decoding-time controlled text generation with experts and anti-experts

A Liu, M Sap, X Lu, S Swayamdipta… - arXiv preprint arXiv …, 2021 - arxiv.org
Despite recent advances in natural language generation, it remains challenging to control
attributes of generated text. We propose DExperts: Decoding-time Experts, a decoding-time …