Optimizing example selection for retrieval-augmented machine translation with translation memories
Retrieval-augmented machine translation leverages examples from a translation memory by retrieving similar instances. These examples are used to condition the predictions of a neural decoder. We aim to improve the upstream retrieval step and consider a fixed downstream edit-based model: the multi-Le...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Retrieval-augmented machine translation leverages examples from a translation
memory by retrieving similar instances. These examples are used to condition
the predictions of a neural decoder. We aim to improve the upstream retrieval
step and consider a fixed downstream edit-based model: the multi-Levenshtein
Transformer. The task consists of finding a set of examples that maximizes the
overall coverage of the source sentence. To this end, we rely on the theory of
submodular functions and explore new algorithms to optimize this coverage. We
evaluate the resulting performance gains for the machine translation task. |
---|---|
DOI: | 10.48550/arxiv.2405.15070 |