A Primer on Pretrained Multilingual Language Models
Multilingual Language Models (\MLLMs) such as mBERT, XLM, XLM-R, \textit{etc.} have emerged as a viable option for bringing the power of pretraining to a large number of languages. Given their success in zero-shot transfer learning, there has emerged a large body of work in (i) building bigger \MLLM...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Multilingual Language Models (\MLLMs) such as mBERT, XLM, XLM-R,
\textit{etc.} have emerged as a viable option for bringing the power of
pretraining to a large number of languages. Given their success in zero-shot
transfer learning, there has emerged a large body of work in (i) building
bigger \MLLMs~covering a large number of languages (ii) creating exhaustive
benchmarks covering a wider variety of tasks and languages for evaluating
\MLLMs~ (iii) analysing the performance of \MLLMs~on monolingual, zero-shot
cross-lingual and bilingual tasks (iv) understanding the universal language
patterns (if any) learnt by \MLLMs~ and (v) augmenting the (often) limited
capacity of \MLLMs~ to improve their performance on seen or even unseen
languages. In this survey, we review the existing literature covering the above
broad areas of research pertaining to \MLLMs. Based on our survey, we recommend
some promising directions of future research. |
---|---|
DOI: | 10.48550/arxiv.2107.00676 |