Pre-trained models for natural language processing: A survey

X Qiu, T Sun, Y Xu, Y Shao, N Dai, X Huang - Science China …, 2020 - Springer
Recently, the emergence of pre-trained models (PTMs) has brought natural language
processing (NLP) to a new era. In this survey, we provide a comprehensive review of PTMs …

Domain adaptation for visual applications: A comprehensive survey

G Csurka - arXiv preprint arXiv:1702.05374, 2017 - arxiv.org
The aim of this paper is to give an overview of domain adaptation and transfer learning with
a specific view on visual applications. After a general motivation, we first position domain …

Xtreme: A massively multilingual multi-task benchmark for evaluating cross-lingual generalisation

J Hu, S Ruder, A Siddhant, G Neubig… - International …, 2020 - proceedings.mlr.press
Much recent progress in applications of machine learning models to NLP has been driven
by benchmarks that evaluate models across a wide variety of tasks. However, these broad …

Cross-lingual language model pretraining

A Conneau, G Lample - Advances in neural information …, 2019 - proceedings.neurips.cc
Recent studies have demonstrated the efficiency of generative pretraining for English
natural language understanding. In this work, we extend this approach to multiple …

XNLI: Evaluating cross-lingual sentence representations

A Conneau, G Lample, R Rinott, A Williams… - arXiv preprint arXiv …, 2018 - arxiv.org
State-of-the-art natural language processing systems rely on supervision in the form of
annotated data to learn competent models. These models are generally trained on data in a …

Word translation without parallel data

A Conneau, G Lample, MA Ranzato, L Denoyer… - arXiv preprint arXiv …, 2017 - arxiv.org
State-of-the-art methods for learning cross-lingual word embeddings have relied on
bilingual dictionaries or parallel corpora. Recent studies showed that the need for parallel …

A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings

M Artetxe, G Labaka, E Agirre - arXiv preprint arXiv:1805.06297, 2018 - arxiv.org
Recent work has managed to learn cross-lingual word embeddings without parallel data by
mapping monolingual embeddings to a shared space through adversarial training …

[图书][B] Neural network methods in natural language processing

Y Goldberg - 2017 - books.google.com
Neural networks are a family of powerful machine learning models and this book focuses on
their application to natural language data. The first half of the book (Parts I and II) covers the …

[PDF][PDF] Word translation without parallel data

G Lample, A Conneau, MA Ranzato… - International …, 2018 - openreview.net
State-of-the-art methods for learning cross-lingual word embeddings have relied on
bilingual dictionaries or parallel corpora. Recent studies showed that the need for parallel …

A survey of cross-lingual word embedding models

S Ruder, I Vulić, A Søgaard - Journal of Artificial Intelligence Research, 2019 - jair.org
Cross-lingual representations of words enable us to reason about word meaning in
multilingual contexts and are a key facilitator of cross-lingual transfer when developing …