Pre-trained language models for text generation: A survey

J Li, T Tang, WX Zhao, JY Nie, JR Wen - ACM Computing Surveys, 2024 - dl.acm.org
Text Generation aims to produce plausible and readable text in human language from input
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …

Virtual customer assistants in finance: From state of the art and practices to design guidelines

A Iovine, F Narducci, C Musto, M de Gemmis… - Computer Science …, 2023 - Elsevier
Abstract Virtual Customer Assistants (VCAs) are revolutionizing the way users interact with
machines. VCAs allow a far more natural interaction, and are gaining an increasingly large …

Conversations are not flat: Modeling the dynamic information flow across dialogue utterances

Z Li, J Zhang, Z Fei, Y Feng, J Zhou - arXiv preprint arXiv:2106.02227, 2021 - arxiv.org
Nowadays, open-domain dialogue models can generate acceptable responses according to
the historical context based on the large-scale pre-trained language models. However, they …

Teacher forcing recovers reward functions for text generation

Y Hao, Y Liu, L Mou - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Reinforcement learning (RL) has been widely used in text generation to alleviate the
exposure bias issue or to utilize non-parallel datasets. The reward function plays an …

Structural pre-training for dialogue comprehension

Z Zhang, H Zhao - arXiv preprint arXiv:2105.10956, 2021 - arxiv.org
Pre-trained language models (PrLMs) have demonstrated superior performance due to their
strong ability to learn universal language representations from self-supervised pre-training …

Section-aware commonsense knowledge-grounded dialogue generation with pre-trained language model

S Wu, Y Li, P Xue, D Zhang, Z Wu - Proceedings of the 29th …, 2022 - aclanthology.org
In knowledge-grounded dialogue generation, pre-trained language models (PLMs) can be
expected to deepen the fusing of dialogue context and knowledge because of their superior …

Task compass: Scaling multi-task pre-training with task prefix

Z Zhang, S Wang, Y Xu, Y Fang, W Yu, Y Liu… - arXiv preprint arXiv …, 2022 - arxiv.org
Leveraging task-aware annotated data as supervised signals to assist with self-supervised
learning on large-scale unlabeled data has become a new trend in pre-training language …

An equal-size hard EM algorithm for diverse dialogue generation

Y Wen, Y Hao, Y Cao, L Mou - arXiv preprint arXiv:2209.14627, 2022 - arxiv.org
Open-domain dialogue systems aim to interact with humans through natural language texts
in an open-ended fashion. Despite the recent success of super large dialogue systems such …

A Primer on Seq2Seq Models for Generative Chatbots

V Scotti, L Sbattella, R Tedesco - ACM Computing Surveys, 2023 - dl.acm.org
The recent spread of Deep Learning-based solutions for Artificial Intelligence and the
development of Large Language Models has pushed forwards significantly the Natural …

Emp-rft: Empathetic response generation via recognizing feature transitions between utterances

W Kim, Y Ahn, D Kim, KH Lee - arXiv preprint arXiv:2205.03112, 2022 - arxiv.org
Each utterance in multi-turn empathetic dialogues has features such as emotion, keywords,
and utterance-level meaning. Feature transitions between utterances occur naturally …