Dart: Open-domain structured data record to text generation

L Nan, D Radev, R Zhang, A Rau, A Sivaprasad… - arXiv preprint arXiv …, 2020 - arxiv.org
We present DART, an open domain structured DAta Record to Text generation dataset with
over 82k instances (DARTs). Data-to-Text annotations can be a costly process, especially …

Barack's wife Hillary: Using knowledge-graphs for fact-aware language modeling

RL Logan IV, NF Liu, ME Peters, M Gardner… - arXiv preprint arXiv …, 2019 - arxiv.org
Modeling human language requires the ability to not only generate fluent text but also
encode factual knowledge. However, traditional language models are only capable of …

Neural data-to-text generation: A comparison between pipeline and end-to-end architectures

TC Ferreira, C van der Lee, E Van Miltenburg… - arXiv preprint arXiv …, 2019 - arxiv.org
Traditionally, most data-to-text applications have been designed using a modular pipeline
architecture, in which non-linguistic input data is converted into natural language through …

A survey on neural data-to-text generation

Y Lin, T Ruan, J Liu, H Wang - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Data-to-text Generation (D2T) aims to generate textual natural language statements that can
fluently and precisely describe the structured data such as graphs, tables, and meaning …

Plan-then-generate: Controlled data-to-text generation via planning

Y Su, D Vandyke, S Wang, Y Fang, N Collier - arXiv preprint arXiv …, 2021 - arxiv.org
Recent developments in neural networks have led to the advance in data-to-text generation.
However, the lack of ability of neural models to control the structure of generated output can …

Bridging the structural gap between encoding and decoding for data-to-text generation

C Zhao, M Walker, S Chaturvedi - … of the 58th annual meeting of …, 2020 - aclanthology.org
Generating sequential natural language descriptions from graph-structured data (eg,
knowledge graph) is challenging, partly because of the structural differences between the …

Have your text and use it too! end-to-end neural data-to-text generation with semantic fidelity

H Harkous, I Groves, A Saffari - arXiv preprint arXiv:2004.06577, 2020 - arxiv.org
End-to-end neural data-to-text (D2T) generation has recently emerged as an alternative to
pipeline-based architectures. However, it has faced challenges in generalizing to new …

Automatic text evaluation through the lens of Wasserstein barycenters

P Colombo, G Staerman, C Clavel… - arXiv preprint arXiv …, 2021 - arxiv.org
A new metric\texttt {BaryScore} to evaluate text generation based on deep contextualized
embeddings eg, BERT, Roberta, ELMo) is introduced. This metric is motivated by a new …

Infolm: A new metric to evaluate summarization & data2text generation

PJA Colombo, C Clavel, P Piantanida - Proceedings of the AAAI …, 2022 - ojs.aaai.org
Assessing the quality of natural language generation (NLG) systems through human
annotation is very expensive. Additionally, human annotation campaigns are time …

Decoder-only or encoder-decoder? interpreting language model as a regularized encoder-decoder

Z Fu, W Lam, Q Yu, AMC So, S Hu, Z Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
The sequence-to-sequence (seq2seq) task aims at generating the target sequence based
on the given input source sequence. Traditionally, most of the seq2seq task is resolved by …