A cost-benefit analysis of cross-lingual transfer methods
An effective method for cross-lingual transfer is to fine-tune a bilingual or multilingual model on a supervised dataset in one language and evaluating it on another language in a zero-shot manner. Translating examples at training time or inference time are also viable alternatives. However, there a...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | An effective method for cross-lingual transfer is to fine-tune a bilingual or
multilingual model on a supervised dataset in one language and evaluating it on
another language in a zero-shot manner. Translating examples at training time
or inference time are also viable alternatives. However, there are costs
associated with these methods that are rarely addressed in the literature. In
this work, we analyze cross-lingual methods in terms of their effectiveness
(e.g., accuracy), development and deployment costs, as well as their latencies
at inference time. Our experiments on three tasks indicate that the best
cross-lingual method is highly task-dependent. Finally, by combining zero-shot
and translation methods, we achieve the state-of-the-art in two of the three
datasets used in this work. Based on these results, we question the need for
manually labeled training data in a target language. Code and translated
datasets are available at https://github.com/unicamp-dl/cross-lingual-analysis |
---|---|
DOI: | 10.48550/arxiv.2105.06813 |