Scigen: a dataset for reasoning-aware text generation from scientific tables

NS Moosavi, A Rücklé, D Roth… - Thirty-fifth Conference on …, 2021 - openreview.net
We introduce SciGen, a new challenge dataset consisting of tables from scientific articles
and their corresponding descriptions, for the task of reasoning-aware data-to-text …

The curse of performance instability in analysis datasets: Consequences, source, and suggestions

X Zhou, Y Nie, H Tan, M Bansal - arXiv preprint arXiv:2004.13606, 2020 - arxiv.org
We find that the performance of state-of-the-art models on Natural Language Inference (NLI)
and Reading Comprehension (RC) analysis/stress sets can be highly unstable. This raises …

Learning to reason for text generation from scientific tables

NS Moosavi, A Rücklé, D Roth, I Gurevych - arXiv preprint arXiv …, 2021 - arxiv.org
In this paper, we introduce SciGen, a new challenge dataset for the task of reasoning-aware
data-to-text generation consisting of tables from scientific articles and their corresponding …

Improving robustness by augmenting training sentences with predicate-argument structures

NS Moosavi, M de Boer, PA Utama… - arXiv preprint arXiv …, 2020 - arxiv.org
Existing NLP datasets contain various biases, and models tend to quickly learn those
biases, which in turn limits their robustness. Existing approaches to improve robustness …

Robustness of Pre-trained Language Models for Natural Language Understanding

PA Utama - 2024 - tuprints.ulb.tu-darmstadt.de
Recent advances in neural network architectures and large-scale language model
pretraining have enabled Natural Language Understanding (NLU) systems to surpass …

[PDF][PDF] Representation learning and learning from limited labeled data for community question answering

A Rücklé - 2021 - tuprints.ulb.tu-darmstadt.de
The amount of information published on the Internet is growing steadily. Accessing the vast
knowledge in them more effectively is a fundamental goal of many tasks in natural language …

Towards Reliable and Generalizable Natural Language Understanding

X Zhou - 2023 - search.proquest.com
The de facto paradigm of developing NLP models requires collecting a dataset, training the
model on the training set assuming all the examples are iid (independent and identically …