A survey of deep learning techniques for neural machine translation

S Yang, Y Wang, X Chu - arXiv preprint arXiv:2002.07526, 2020 - arxiv.org
In recent years, natural language processing (NLP) has got great development with deep
learning techniques. In the sub-field of machine translation, a new approach named Neural …

Towards robust neural machine translation

Y Cheng, Z Tu, F Meng, J Zhai, Y Liu - arXiv preprint arXiv:1805.06130, 2018 - arxiv.org
Small perturbations in the input can severely distort intermediate representations and thus
impact translation quality of neural machine translation (NMT) models. In this paper, we …

Neural machine translation with GRU-gated attention model

B Zhang, D Xiong, J Xie, J Su - IEEE transactions on neural …, 2020 - ieeexplore.ieee.org
Neural machine translation (NMT) heavily relies on context vectors generated by an
attention network to predict target words. In practice, we observe that the context vectors for …

Training deeper neural machine translation models with transparent attention

A Bapna, MX Chen, O Firat, Y Cao, Y Wu - arXiv preprint arXiv:1808.07561, 2018 - arxiv.org
While current state-of-the-art NMT models, such as RNN seq2seq and Transformers,
possess a large number of parameters, they are still shallow in comparison to convolutional …

Neural machine translation with deep attention

B Zhang, D Xiong, J Su - IEEE transactions on pattern analysis …, 2018 - ieeexplore.ieee.org
Deepening neural models has been proven very successful in improving the model's
capacity when solving complex learning tasks, such as the machine translation task …

Improving deep transformer with depth-scaled initialization and merged attention

B Zhang, I Titov, R Sennrich - arXiv preprint arXiv:1908.11365, 2019 - arxiv.org
The general trend in NLP is towards increasing model capacity and performance via deeper
neural networks. However, simply stacking more layers of the popular Transformer …

Asynchronous bidirectional decoding for neural machine translation

X Zhang, J Su, Y Qin, Y Liu, R Ji, H Wang - Proceedings of the AAAI …, 2018 - ojs.aaai.org
The dominant neural machine translation (NMT) models apply unified attentional encoder-
decoder neural networks for translation. Traditionally, the NMT decoders adopt recurrent …

Graph-to-tree neural networks for learning structured input-output translation with applications to semantic parsing and math word problem

S Li, L Wu, S Feng, F Xu, F Xu, S Zhong - arXiv preprint arXiv:2004.13781, 2020 - arxiv.org
The celebrated Seq2Seq technique and its numerous variants achieve excellent
performance on many tasks such as neural machine translation, semantic parsing, and math …

Variational recurrent neural machine translation

J Su, S Wu, D Xiong, Y Lu, X Han… - Proceedings of the AAAI …, 2018 - ojs.aaai.org
Partially inspired by successful applications of variational recurrent neural networks, we
propose a novel variational recurrent neural machine translation (VRNMT) model in this …

Regularizing neural machine translation by target-bidirectional agreement

Z Zhang, S Wu, S Liu, M Li, M Zhou, T Xu - Proceedings of the AAAI …, 2019 - aaai.org
Abstract Although Neural Machine Translation (NMT) has achieved remarkable progress in
the past several years, most NMT systems still suffer from a fundamental shortcoming as in …