Self-attention with cross-lingual position representation

L Ding, L Wang, D Tao - arXiv preprint arXiv:2004.13310, 2020 - arxiv.org
Position encoding (PE), an essential part of self-attention networks (SANs), is used to
preserve the word order information for natural language processing tasks, generating fixed …

Guiding non-autoregressive neural machine translation decoding with reordering information

Q Ran, Y Lin, P Li, J Zhou - Proceedings of the AAAI Conference on …, 2021 - ojs.aaai.org
Non-autoregressive neural machine translation (NAT) generates each target word in parallel
and has achieved promising inference acceleration. However, existing NAT models still …

Structural pre-training for dialogue comprehension

Z Zhang, H Zhao - arXiv preprint arXiv:2105.10956, 2021 - arxiv.org
Pre-trained language models (PrLMs) have demonstrated superior performance due to their
strong ability to learn universal language representations from self-supervised pre-training …

Bilingual attention based neural machine translation

L Kang, S He, M Wang, F Long, J Su - Applied Intelligence, 2023 - Springer
Abstract In recent years, Recurrent Neural Network based Neural Machine Translation (RNN-
based NMT) equipped with an attention mechanism from the decoder to encoder, has …

Structured reordering for modeling latent alignments in sequence transduction

M Lapata, I Titov - Advances in Neural Information …, 2021 - proceedings.neurips.cc
Despite success in many domains, neural models struggle in settings where train and test
examples are drawn from different distributions. In particular, in contrast to humans …

A hierarchical clustering approach to fuzzy semantic representation of rare words in neural machine translation

M Yang, S Liu, K Chen, H Zhang… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
Rare words are usually replaced with a single token in the current encoder–decoder style of
neural machine translation, challenging the translation modeling by an obscured context. In …

[PDF][PDF] 神经机器翻译前沿综述

冯洋, 邵晨泽 - 中文信息学报, 2020 - jcip.cipsc.org.cn
机器翻译是指通过计算机将源语言句子翻译到与之语义等价的目标语言句子的过程,
是自然语言处理领域的一个重要研究方向. 神经机器翻译仅需使用神经网络就能实现从源语言到 …

Towards enhancing faithfulness for neural machine translation

R Weng, H Yu, X Wei, W Luo - Proceedings of the 2020 …, 2020 - aclanthology.org
Neural machine translation (NMT) has achieved great success due to the ability to generate
high-quality sentences. Compared with human translations, one of the drawbacks of current …

Recurrent positional embedding for neural machine translation

K Chen, R Wang, M Utiyama… - Proceedings of the 2019 …, 2019 - aclanthology.org
In the Transformer network architecture, positional embeddings are used to encode order
dependencies into the input representation. However, this input representation only involves …

Multi-head highly parallelized LSTM decoder for neural machine translation

H Xu, Q Liu, J van Genabith, D Xiong… - Proceedings of the 59th …, 2021 - aclanthology.org
One of the reasons Transformer translation models are popular is that self-attention
networks for context modelling can be easily parallelized at sequence level. However, the …