Multilingual Transformer Language Model for Speech Recognition in Low-resource Languages
It is challenging to train and deploy Transformer LMs for hybrid speech recognition 2nd pass re-ranking in low-resource languages due to (1) data scarcity in low-resource languages, (2) expensive computing costs for training and refreshing 100+ monolingual models, and (3) hosting inefficiency consid...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | It is challenging to train and deploy Transformer LMs for hybrid speech
recognition 2nd pass re-ranking in low-resource languages due to (1) data
scarcity in low-resource languages, (2) expensive computing costs for training
and refreshing 100+ monolingual models, and (3) hosting inefficiency
considering sparse traffic. In this study, we present a new way to group
multiple low-resource locales together and optimize the performance of
Multilingual Transformer LMs in ASR. Our Locale-group Multilingual Transformer
LMs outperform traditional multilingual LMs along with reducing maintenance
costs and operating expenses. Further, for low-resource but high-traffic
locales where deploying monolingual models is feasible, we show that
fine-tuning our locale-group multilingual LMs produces better monolingual LM
candidates than baseline monolingual LMs. |
---|---|
DOI: | 10.48550/arxiv.2209.04041 |