[HTML][HTML] Data augmentation approaches in natural language processing: A survey

B Li, Y Hou, W Che - Ai Open, 2022 - Elsevier
As an effective strategy, data augmentation (DA) alleviates data scarcity scenarios where
deep learning techniques may fail. It is widely applied in computer vision then introduced to …

A survey of data augmentation approaches for NLP

SY Feng, V Gangal, J Wei, S Chandar… - arXiv preprint arXiv …, 2021 - arxiv.org
Data augmentation has recently seen increased interest in NLP due to more work in low-
resource domains, new tasks, and the popularity of large-scale neural networks that require …

Gpt3mix: Leveraging large-scale language models for text augmentation

KM Yoo, D Park, J Kang, SW Lee, W Park - arXiv preprint arXiv …, 2021 - arxiv.org
Large-scale language models such as GPT-3 are excellent few-shot learners, allowing them
to be controlled via natural text prompts. Recent studies report that prompt-based direct …

Spectrum interference-based two-level data augmentation method in deep learning for automatic modulation classification

Q Zheng, P Zhao, Y Li, H Wang, Y Yang - Neural Computing and …, 2021 - Springer
Automatic modulation classification is an essential and challenging topic in the development
of cognitive radios, and it is the cornerstone of adaptive modulation and demodulation …

Increasing diversity while maintaining accuracy: Text data generation with large language models and human interventions

JJY Chung, E Kamar, S Amershi - arXiv preprint arXiv:2306.04140, 2023 - arxiv.org
Large language models (LLMs) can be used to generate text data for training and evaluating
other models. However, creating high-quality datasets with LLMs can be challenging. In this …

An analysis of simple data augmentation for named entity recognition

X Dai, H Adel - arXiv preprint arXiv:2010.11683, 2020 - arxiv.org
Simple yet effective data augmentation techniques have been proposed for sentence-level
and sentence-pair natural language processing tasks. Inspired by these efforts, we design …

Few-shot slot tagging with collapsed dependency transfer and label-enhanced task-adaptive projection network

Y Hou, W Che, Y Lai, Z Zhou, Y Liu, H Liu… - arXiv preprint arXiv …, 2020 - arxiv.org
In this paper, we explore the slot tagging with only a few labeled support sentences (aka few-
shot). Few-shot slot tagging faces a unique challenge compared to the other few-shot …

Mixup-transformer: Dynamic data augmentation for NLP tasks

L Sun, C Xia, W Yin, T Liang, PS Yu, L He - arXiv preprint arXiv …, 2020 - arxiv.org
Mixup is the latest data augmentation technique that linearly interpolates input examples
and the corresponding labels. It has shown strong effectiveness in image classification by …

Recent neural methods on slot filling and intent classification for task-oriented dialogue systems: A survey

S Louvan, B Magnini - arXiv preprint arXiv:2011.00564, 2020 - arxiv.org
In recent years, fostered by deep learning technologies and by the high demand for
conversational AI, various approaches have been proposed that address the capacity to …

Flipda: Effective and robust data augmentation for few-shot learning

J Zhou, Y Zheng, J Tang, J Li, Z Yang - arXiv preprint arXiv:2108.06332, 2021 - arxiv.org
Most previous methods for text data augmentation are limited to simple tasks and weak
baselines. We explore data augmentation on hard tasks (ie, few-shot natural language …