Prompting Large Language Models with Human Error Markings for Self-Correcting Machine Translation

N Berger, S Riezler, M Exel, M Huck - arXiv preprint arXiv:2406.02267, 2024 - arxiv.org
While large language models (LLMs) pre-trained on massive amounts of unpaired language
data have reached the state-of-the-art in machine translation (MT) of general domain texts …

Post-edits Are Preferences Too

N Berger, S Riezler, M Exel, M Huck - arXiv preprint arXiv:2410.02320, 2024 - arxiv.org
Preference Optimization (PO) techniques are currently one of the state of the art techniques
for fine-tuning large language models (LLMs) on pairwise preference feedback from human …