An empirical survey on long document summarization: Datasets, models, and metrics

HY Koh, J Ju, M Liu, S Pan - ACM computing surveys, 2022 - dl.acm.org
Long documents such as academic articles and business reports have been the standard
format to detail out important issues and complicated subjects that require extra attention. An …

Multi-XScience: A large-scale dataset for extreme multi-document summarization of scientific articles

Y Lu, Y Dong, L Charlin - arXiv preprint arXiv:2010.14235, 2020 - arxiv.org
Multi-document summarization is a challenging task for which there exists little large-scale
datasets. We propose Multi-XScience, a large-scale multi-document summarization dataset …

Does pretraining for summarization require knowledge transfer?

K Krishna, J Bigham, ZC Lipton - arXiv preprint arXiv:2109.04953, 2021 - arxiv.org
Pretraining techniques leveraging enormous datasets have driven recent advances in text
summarization. While folk explanations suggest that knowledge transfer accounts for …

Discourse-aware unsupervised summarization of long scientific documents

Y Dong, A Mircea, JCK Cheung - arXiv preprint arXiv:2005.00513, 2020 - arxiv.org
We propose an unsupervised graph-based ranking model for extractive summarization of
long scientific documents. Our method assumes a two-level hierarchical graph …

StreamHover: Livestream transcript summarization and annotation

S Cho, F Dernoncourt, T Ganter, T Bui, N Lipka… - arXiv preprint arXiv …, 2021 - arxiv.org
With the explosive growth of livestream broadcasting, there is an urgent need for new
summarization technology that enables us to create a preview of streamed content and tap …

Revisiting zero-shot abstractive summarization in the era of large language models from the perspective of position bias

A Chhabra, H Askari, P Mohapatra - arXiv preprint arXiv:2401.01989, 2024 - arxiv.org
We characterize and study zero-shot abstractive summarization in Large Language Models
(LLMs) by measuring position bias, which we propose as a general formulation of the more …

Ffci: A framework for interpretable automatic evaluation of summarization

F Koto, T Baldwin, JH Lau - Journal of Artificial Intelligence Research, 2022 - jair.org
In this paper, we propose FFCI, a framework for fine-grained summarization evaluation that
comprises four elements: faithfulness (degree of factual consistency with the source), focus …

Leveraging lead bias for zero-shot abstractive news summarization

C Zhu, Z Yang, R Gmyr, M Zeng, X Huang - Proceedings of the 44th …, 2021 - dl.acm.org
A typical journalistic convention in news articles is to deliver the most salient information in
the beginning, also known as the lead bias. While this phenomenon can be exploited in …

A sliding-window approach to automatic creation of meeting minutes

JJ Koay, A Roustai, X Dai, F Liu - arXiv preprint arXiv:2104.12324, 2021 - arxiv.org
Meeting minutes record any subject matters discussed, decisions reached and actions taken
at meetings. The importance of minuting cannot be overemphasized in a time when a …

Demoting the lead bias in news summarization via alternating adversarial learning

L Xing, W Xiao, G Carenini - arXiv preprint arXiv:2105.14241, 2021 - arxiv.org
In news articles the lead bias is a common phenomenon that usually dominates the learning
signals for neural extractive summarizers, severely limiting their performance on data with …