Q Ran, Y Lin, P Li, J Zhou - Proceedings of the AAAI Conference on …, 2021 - ojs.aaai.org
Non-autoregressive neural machine translation (NAT) generates each target word in parallel and has achieved promising inference acceleration. However, existing NAT models still …
Z Zhang, H Zhao - arXiv preprint arXiv:2105.10956, 2021 - arxiv.org
Pre-trained language models (PrLMs) have demonstrated superior performance due to their strong ability to learn universal language representations from self-supervised pre-training …
L Kang, S He, M Wang, F Long, J Su - Applied Intelligence, 2023 - Springer
Abstract In recent years, Recurrent Neural Network based Neural Machine Translation (RNN- based NMT) equipped with an attention mechanism from the decoder to encoder, has …
M Lapata, I Titov - Advances in Neural Information …, 2021 - proceedings.neurips.cc
Despite success in many domains, neural models struggle in settings where train and test examples are drawn from different distributions. In particular, in contrast to humans …
M Yang, S Liu, K Chen, H Zhang… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
Rare words are usually replaced with a single token in the current encoder–decoder style of neural machine translation, challenging the translation modeling by an obscured context. In …
Neural machine translation (NMT) has achieved great success due to the ability to generate high-quality sentences. Compared with human translations, one of the drawbacks of current …
In the Transformer network architecture, positional embeddings are used to encode order dependencies into the input representation. However, this input representation only involves …
One of the reasons Transformer translation models are popular is that self-attention networks for context modelling can be easily parallelized at sequence level. However, the …