Mutual information alleviates hallucinations in abstractive summarization

L Van der Poel, R Cotterell, C Meister - arXiv preprint arXiv:2210.13210, 2022 - arxiv.org
Despite significant progress in the quality of language generated from abstractive
summarization models, these models still exhibit the tendency to hallucinate, ie, output …

Don't say what you don't know: Improving the consistency of abstractive summarization by constraining beam search

D King, Z Shen, N Subramani, DS Weld… - arXiv preprint arXiv …, 2022 - arxiv.org
Abstractive summarization systems today produce fluent and relevant output, but often"
hallucinate" statements not supported by the source text. We analyze the connection …

Learning with rejection for abstractive text summarization

M Cao, Y Dong, J He, JCK Cheung - arXiv preprint arXiv:2302.08531, 2023 - arxiv.org
State-of-the-art abstractive summarization systems frequently hallucinate content that is not
supported by the source document, mainly due to noise in the training dataset. Existing …

Training dynamics for text summarization models

T Goyal, J Xu, JJ Li, G Durrett - arXiv preprint arXiv:2110.08370, 2021 - arxiv.org
Pre-trained language models (eg BART) have shown impressive results when fine-tuned on
large summarization datasets. However, little is understood about this fine-tuning process …

Constrained abstractive summarization: Preserving factual consistency with constrained generation

Y Mao, X Ren, H Ji, J Han - arXiv preprint arXiv:2010.12723, 2020 - arxiv.org
Despite significant progress, state-of-the-art abstractive summarization methods are still
prone to hallucinate content inconsistent with the source document. In this paper, we …

Faithfulness-aware decoding strategies for abstractive summarization

D Wan, M Liu, K McKeown, M Dreyer… - arXiv preprint arXiv …, 2023 - arxiv.org
Despite significant progress in understanding and improving faithfulness in abstractive
summarization, the question of how decoding strategies affect faithfulness is less studied …

Alleviating exposure bias via contrastive learning for abstractive text summarization

S Sun, W Li - arXiv preprint arXiv:2108.11846, 2021 - arxiv.org
Encoder-decoder models have achieved remarkable success in abstractive text
summarization, which aims to compress one or more documents into a shorter version …

Understanding neural abstractive summarization models via uncertainty

J Xu, S Desai, G Durrett - arXiv preprint arXiv:2010.07882, 2020 - arxiv.org
An advantage of seq2seq abstractive summarization models is that they generate text in a
free-form manner, but this flexibility makes it difficult to interpret model behavior. In this work …

Diversity driven attention model for query-based abstractive summarization

P Nema, M Khapra, A Laha, B Ravindran - arXiv preprint arXiv …, 2017 - arxiv.org
Abstractive summarization aims to generate a shorter version of the document covering all
the salient points in a compact and coherent fashion. On the other hand, query-based …

Length control in abstractive summarization by pretraining information selection

Y Liu, Q Jia, K Zhu - Proceedings of the 60th Annual Meeting of …, 2022 - aclanthology.org
Previous length-controllable summarization models mostly control lengths at the decoding
stage, whereas the encoding or the selection of information from the source document is not …