Chestxraybert: A pretrained language model for chest radiology report summarization

X Cai, S Liu, J Han, L Yang, Z Liu… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Automatically generating the “impression” section of a radiology report given the “findings”
section can summarize as much salient information of the “findings” section as possible, thus …

Neural data-to-text generation with LM-based text augmentation

E Chang, X Shen, D Zhu, V Demberg, H Su - arXiv preprint arXiv …, 2021 - arxiv.org
For many new application domains for data-to-text generation, the main obstacle in training
neural models consists of a lack of training data. While usually large numbers of instances …

On training instance selection for few-shot neural text generation

E Chang, X Shen, HS Yeh, V Demberg - arXiv preprint arXiv:2107.03176, 2021 - arxiv.org
Large-scale pretrained language models have led to dramatic improvements in text
generation. Impressive performance can be achieved by finetuning only on a small number …

Does the order of training samples matter? improving neural data-to-text generation with curriculum learning

E Chang, HS Yeh, V Demberg - arXiv preprint arXiv:2102.03554, 2021 - arxiv.org
Recent advancements in data-to-text generation largely take on the form of neural end-to-
end systems. Efforts have been dedicated to improving text generation systems by changing …

Weakly supervised spatial relation extraction from radiology reports

S Datta, K Roberts - JAMIA open, 2023 - academic.oup.com
Objective Weak supervision holds significant promise to improve clinical natural language
processing by leveraging domain resources and expertise instead of large manually …

[HTML][HTML] Language as a latent sequence: Deep latent variable models for semi-supervised paraphrase generation

J Yu, AI Cristea, A Harit, Z Sun, OT Aduragba, L Shi… - AI Open, 2023 - Elsevier
This paper explores deep latent variable models for semi-supervised paraphrase
generation, where the missing target pair for unlabelled data is modelled as a latent …

ReMeDi: Resources for multi-domain, multi-service, medical dialogues

G Yan, J Pei, P Ren, Z Ren, X Xin, H Liang… - Proceedings of the 45th …, 2022 - dl.acm.org
\AcpMDS aim to assist doctors and patients with a range of professional medical services, ie,
diagnosis, treatment and consultation. The development of\acpMDS is hindered because of …

[PDF][PDF] Neural data-to-text generation based on small datasets: Comparing the added value of two semi-supervised learning approaches on top of a large language …

C van der Lee, TC Ferreira, C Emmery… - Computational …, 2023 - direct.mit.edu
This study discusses the effect of semi-supervised learning in combination with pretrained
language models for data-to-text generation. It is not known whether semi-supervised …

Unsupervised pidgin text generation by pivoting english data and self-training

E Chang, DI Adelani, X Shen, V Demberg - arXiv preprint arXiv …, 2020 - arxiv.org
West African Pidgin English is a language that is significantly spoken in West Africa,
consisting of at least 75 million speakers. Nevertheless, proper machine translation systems …

Contextual knowledge learning for dialogue generation

W Zheng, N Milic-Frayling, K Zhou - arXiv preprint arXiv:2305.18200, 2023 - arxiv.org
Incorporating conversational context and knowledge into dialogue generation models has
been essential for improving the quality of the generated responses. The context, comprising …