Contrastive conditioning for assessing disambiguation in MT: A case study of distilled bias

J Vamvas, R Sennrich - 2021 Conference on Empirical Methods …, 2021 - research.ed.ac.uk
2021 Conference on Empirical Methods in Natural Language Processing, 2021research.ed.ac.uk
Lexical disambiguation is a major challenge for machine translation systems, especially if
some senses of a word are trained less often than others. Identifying patterns of
overgeneralization requires evaluation methods that are both reliable and scalable. We
propose contrastive conditioning as a reference-free blackbox method for detecting
disambiguation errors. Specifically, we score the quality of a translation by conditioning on
variants of the source that provide contrastive disambiguation cues. After validating our …
Abstract
Lexical disambiguation is a major challenge for machine translation systems, especially if some senses of a word are trained less often than others. Identifying patterns of overgeneralization requires evaluation methods that are both reliable and scalable. We propose contrastive conditioning as a reference-free blackbox method for detecting disambiguation errors. Specifically, we score the quality of a translation by conditioning on variants of the source that provide contrastive disambiguation cues. After validating our method, we apply it in a case study to perform a targeted evaluation of sequence-level knowledge distillation. By probing word sense disambiguation and translation of gendered occupation names, we show that distillation-trained models tend to overgeneralize more than other models with a comparable BLEU score. Contrastive conditioning thus highlights a side effect of distillation that is not fully captured by standard evaluation metrics. Code and data to reproduce our findings are publicly available. 1
research.ed.ac.uk
以上显示的是最相近的搜索结果。 查看全部搜索结果