作者
Weiyue Wang, Zijian Yang, Yingbo Gao, Hermann Ney
发表日期
2021/8
研讨会论文
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing: Student Research Workshop
页码范围
23-32
简介
The neural hidden Markov model has been proposed as an alternative to attention mechanism in machine translation with recurrent neural networks. However, since the introduction of the transformer models, its performance has been surpassed. This work proposes to introduce the concept of the hidden Markov model to the transformer architecture, which outperforms the transformer baseline. Interestingly, we find that the zero-order model already provides promising performance, giving it an edge compared to a model with first-order dependency, which performs similarly but is significantly slower in training and decoding.
引用总数
学术搜索中的文章
W Wang, Z Yang, Y Gao, H Ney - Proceedings of the 59th Annual Meeting of the …, 2021