A survey on deep semi-supervised learning

X Yang, Z Song, I King, Z Xu - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Deep semi-supervised learning is a fast-growing field with a range of practical applications.
This paper provides a comprehensive survey on both fundamentals and recent advances in …

A review on question generation from natural language text

R Zhang, J Guo, L Chen, Y Fan, X Cheng - ACM Transactions on …, 2021 - dl.acm.org
Question generation is an important yet challenging problem in Artificial Intelligence (AI),
which aims to generate natural and relevant questions from various input formats, eg …

Minilm: Deep self-attention distillation for task-agnostic compression of pre-trained transformers

W Wang, F Wei, L Dong, H Bao… - Advances in Neural …, 2020 - proceedings.neurips.cc
Pre-trained language models (eg, BERT (Devlin et al., 2018) and its variants) have achieved
remarkable success in varieties of NLP tasks. However, these models usually consist of …

Unified language model pre-training for natural language understanding and generation

L Dong, N Yang, W Wang, F Wei… - Advances in neural …, 2019 - proceedings.neurips.cc
This paper presents a new Unified pre-trained Language Model (UniLM) that can be fine-
tuned for both natural language understanding and generation tasks. The model is pre …

Matching structure for dual learning

H Fei, S Wu, Y Ren, M Zhang - international conference on …, 2022 - proceedings.mlr.press
Many natural language processing (NLP) tasks appear in dual forms, which are generally
solved by dual learning technique that models the dualities between the coupled tasks. In …

Prophetnet: Predicting future n-gram for sequence-to-sequence pre-training

W Qi, Y Yan, Y Gong, D Liu, N Duan, J Chen… - arXiv preprint arXiv …, 2020 - arxiv.org
This paper presents a new sequence-to-sequence pre-training model called ProphetNet,
which introduces a novel self-supervised objective named future n-gram prediction and the …

Unilmv2: Pseudo-masked language models for unified language model pre-training

H Bao, L Dong, F Wei, W Wang… - International …, 2020 - proceedings.mlr.press
We propose to pre-train a unified language model for both autoencoding and partially
autoregressive language modeling tasks using a novel training procedure, referred to as a …

Retrieval-augmented generation for knowledge-intensive nlp tasks

P Lewis, E Perez, A Piktus, F Petroni… - Advances in …, 2020 - proceedings.neurips.cc
Large pre-trained language models have been shown to store factual knowledge in their
parameters, and achieve state-of-the-art results when fine-tuned on downstream NLP tasks …

An empirical survey of data augmentation for limited data learning in nlp

J Chen, D Tam, C Raffel, M Bansal… - Transactions of the …, 2023 - direct.mit.edu
NLP has achieved great progress in the past decade through the use of neural models and
large labeled datasets. The dependence on abundant data prevents NLP models from being …

Evaluating factuality in generation with dependency-level entailment

T Goyal, G Durrett - arXiv preprint arXiv:2010.05478, 2020 - arxiv.org
Despite significant progress in text generation models, a serious limitation is their tendency
to produce text that is factually inconsistent with information in the input. Recent work has …