Active retrieval augmented generation

Z Jiang, FF Xu, L Gao, Z Sun, Q Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
Despite the remarkable ability of large language models (LMs) to comprehend and generate
language, they have a tendency to hallucinate and create factually inaccurate output …

WikiHowQA: A comprehensive benchmark for multi-document non-factoid question answering

V Bolotova-Baranova, V Blinov… - Proceedings of the …, 2023 - aclanthology.org
Answering non-factoid questions (NFQA) is a challenging task, requiring passage-level
answers that are difficult to construct and evaluate. Search engines may provide a summary …

Towards explainable search results: a listwise explanation generator

P Yu, R Rahimi, J Allan - Proceedings of the 45th International ACM …, 2022 - dl.acm.org
It has been shown that the interpretability of search results is enhanced when query aspects
covered by documents are explicitly provided. However, existing work on aspect-oriented …

ROUGE-SEM: Better evaluation of summarization using ROUGE combined with semantics

M Zhang, C Li, M Wan, X Zhang, Q Zhao - Expert Systems with Applications, 2024 - Elsevier
With the development of pre-trained language models and large-scale datasets, automatic
text summarization has attracted much attention from the community of natural language …

WikiDes: A Wikipedia-based dataset for generating short descriptions from paragraphs

HT Ta, ABS Rahman, N Majumder, A Hussain, L Najjar… - Information …, 2023 - Elsevier
As free online encyclopedias with massive volumes of content, Wikipedia and Wikidata are
key to many Natural Language Processing (NLP) tasks, such as information retrieval …

From task to evaluation: an automatic text summarization review

L Lu, Y Liu, W Xu, H Li, G Sun - Artificial Intelligence Review, 2023 - Springer
Automatic summarization is attracting increasing attention as one of the most promising
research areas. This technology has been tried in various real-world applications in recent …

reStructured Pre-training

W Yuan, P Liu - arXiv preprint arXiv:2206.11147, 2022 - arxiv.org
In this work, we try to decipher the internal connection of NLP technology development in the
past decades, searching for essence, which rewards us with a (potential) new learning …

Understanding domain learning in language models through subpopulation analysis

Z Zhao, Y Ziser, SB Cohen - arXiv preprint arXiv:2210.12553, 2022 - arxiv.org
We investigate how different domains are encoded in modern neural network architectures.
We analyze the relationship between natural language domains, model size, and the …

A multi-task learning approach for summarization of dialogues

S Bhattacharjee, K Shinde, T Ghosal… - Proceedings of the 15th …, 2022 - aclanthology.org
We describe our multi-task learning based ap-proach for summarization of real-life
dialogues as part of the DialogSum Challenge shared task at INLG 2022. Our approach …

The State and Fate of Summarization Datasets

N Dahan, G Stanovsky - arXiv preprint arXiv:2411.04585, 2024 - arxiv.org
Automatic summarization has consistently attracted attention, due to its versatility and wide
application in various downstream tasks. Despite its popularity, we find that annotation …