Decoding Time Lexical Domain Adaptationfor Neural Machine Translation
@article{Bogoychev2021DecodingTL, title={Decoding Time Lexical Domain Adaptationfor Neural Machine Translation}, author={Nikolay Bogoychev and Pinzhen Chen}, journal={ArXiv}, year={2021}, volume={abs/2101.00421} }
Machine translation systems are vulnerable to domain mismatch, especially when the task is low-resource. In this setting, out of domain translations are often of poor quality and prone to hallucinations, due to the translation model preferring to predict common words it has seen during training, as opposed to the more uncommon ones from a different domain. We present two simple methods for improving translation quality in this particular setting: First, we use lexical shortlisting in order to… CONTINUE READING
References
SHOWING 1-10 OF 34 REFERENCES
Microsoft Translator at WMT 2019: Towards Large-Scale Document-Level Neural Machine Translation
- Computer Science
- WMT
- 2019
- 54
- Highly Influential
- PDF
Bridging the Gap between Training and Inference for Neural Machine Translation
- Computer Science, Mathematics
- ACL
- 2019
- 63
- PDF
Neural Machine Translation by Minimising the Bayes-risk with Respect to Syntactic Translation Lattices
- Computer Science
- EACL
- 2017
- 26
- PDF