Y Cheng, Z Tu, F Meng, J Zhai, Y Liu - arXiv preprint arXiv:1805.06130, 2018 - arxiv.org
Small perturbations in the input can severely distort intermediate representations and thus impact translation quality of neural machine translation (NMT) models. In this paper, we …
Neural machine translation (NMT) heavily relies on context vectors generated by an attention network to predict target words. In practice, we observe that the context vectors for …
While current state-of-the-art NMT models, such as RNN seq2seq and Transformers, possess a large number of parameters, they are still shallow in comparison to convolutional …
B Zhang, D Xiong, J Su - IEEE transactions on pattern analysis …, 2018 - ieeexplore.ieee.org
Deepening neural models has been proven very successful in improving the model's capacity when solving complex learning tasks, such as the machine translation task …
The general trend in NLP is towards increasing model capacity and performance via deeper neural networks. However, simply stacking more layers of the popular Transformer …
S Li, L Wu, S Feng, F Xu, F Xu, S Zhong - arXiv preprint arXiv:2004.13781, 2020 - arxiv.org
The celebrated Seq2Seq technique and its numerous variants achieve excellent performance on many tasks such as neural machine translation, semantic parsing, and math …
Partially inspired by successful applications of variational recurrent neural networks, we propose a novel variational recurrent neural machine translation (VRNMT) model in this …
Abstract Although Neural Machine Translation (NMT) has achieved remarkable progress in the past several years, most NMT systems still suffer from a fundamental shortcoming as in …