Discovering Representation Sprachbund For Multilingual Pre-Training
Multilingual pre-trained models have demonstrated their effectiveness in many multilingual NLP tasks and enabled zero-shot or few-shot transfer from high-resource languages to low resource ones. However, due to significant typological differences and contradictions between some languages, such model...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Fan, Yimin Liang, Yaobo Muzio, Alexandre Hassan, Hany Li, Houqiang Zhou, Ming Duan, Nan |
description | Multilingual pre-trained models have demonstrated their effectiveness in many
multilingual NLP tasks and enabled zero-shot or few-shot transfer from
high-resource languages to low resource ones. However, due to significant
typological differences and contradictions between some languages, such models
usually perform poorly on many languages and cross-lingual settings, which
shows the difficulty of learning a single model to handle massive diverse
languages well at the same time. To alleviate this issue, we present a new
multilingual pre-training pipeline. We propose to generate language
representation from multilingual pre-trained models and conduct linguistic
analysis to show that language representation similarity reflect linguistic
similarity from multiple perspectives, including language family, geographical
sprachbund, lexicostatistics and syntax. Then we cluster all the target
languages into multiple groups and name each group as a representation
sprachbund. Thus, languages in the same representation sprachbund are supposed
to boost each other in both pre-training and fine-tuning as they share rich
linguistic similarity. We pre-train one multilingual model for each
representation sprachbund. Experiments are conducted on cross-lingual
benchmarks and significant improvements are achieved compared to strong
baselines. |
doi_str_mv | 10.48550/arxiv.2109.00271 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2109_00271</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2109_00271</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-353f934e2d81c2a4e5fa7e02fbaaa4fb211bb3e98707be40a4086ff4a9e28cc23</originalsourceid><addsrcrecordid>eNotz81KxDAUBeBsXMjoA7gyL9B689NJupTqqDCDot2Xm8yNBmpa0nbQt3ccXR0OHA58jF0JKLWtKrjB_BUPpRRQlwDSiHPW3MXJDwfKMb3zVxozTZRmnOOQ-NuY0X-4Je35Zsh8t_Rz7I-7BXv-kqloM8Z07BfsLGA_0eV_rli7uW-bx2L7_PDU3G4LXBtRqEqFWmmSeyu8RE1VQEMgg0NEHZwUwjlFtTVgHGlADXYdgsaapPVeqhW7_rs9Kboxx0_M392vpjtp1A-aS0YM</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Discovering Representation Sprachbund For Multilingual Pre-Training</title><source>arXiv.org</source><creator>Fan, Yimin ; Liang, Yaobo ; Muzio, Alexandre ; Hassan, Hany ; Li, Houqiang ; Zhou, Ming ; Duan, Nan</creator><creatorcontrib>Fan, Yimin ; Liang, Yaobo ; Muzio, Alexandre ; Hassan, Hany ; Li, Houqiang ; Zhou, Ming ; Duan, Nan</creatorcontrib><description>Multilingual pre-trained models have demonstrated their effectiveness in many
multilingual NLP tasks and enabled zero-shot or few-shot transfer from
high-resource languages to low resource ones. However, due to significant
typological differences and contradictions between some languages, such models
usually perform poorly on many languages and cross-lingual settings, which
shows the difficulty of learning a single model to handle massive diverse
languages well at the same time. To alleviate this issue, we present a new
multilingual pre-training pipeline. We propose to generate language
representation from multilingual pre-trained models and conduct linguistic
analysis to show that language representation similarity reflect linguistic
similarity from multiple perspectives, including language family, geographical
sprachbund, lexicostatistics and syntax. Then we cluster all the target
languages into multiple groups and name each group as a representation
sprachbund. Thus, languages in the same representation sprachbund are supposed
to boost each other in both pre-training and fine-tuning as they share rich
linguistic similarity. We pre-train one multilingual model for each
representation sprachbund. Experiments are conducted on cross-lingual
benchmarks and significant improvements are achieved compared to strong
baselines.</description><identifier>DOI: 10.48550/arxiv.2109.00271</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2021-09</creationdate><rights>http://creativecommons.org/licenses/by-sa/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2109.00271$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2109.00271$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Fan, Yimin</creatorcontrib><creatorcontrib>Liang, Yaobo</creatorcontrib><creatorcontrib>Muzio, Alexandre</creatorcontrib><creatorcontrib>Hassan, Hany</creatorcontrib><creatorcontrib>Li, Houqiang</creatorcontrib><creatorcontrib>Zhou, Ming</creatorcontrib><creatorcontrib>Duan, Nan</creatorcontrib><title>Discovering Representation Sprachbund For Multilingual Pre-Training</title><description>Multilingual pre-trained models have demonstrated their effectiveness in many
multilingual NLP tasks and enabled zero-shot or few-shot transfer from
high-resource languages to low resource ones. However, due to significant
typological differences and contradictions between some languages, such models
usually perform poorly on many languages and cross-lingual settings, which
shows the difficulty of learning a single model to handle massive diverse
languages well at the same time. To alleviate this issue, we present a new
multilingual pre-training pipeline. We propose to generate language
representation from multilingual pre-trained models and conduct linguistic
analysis to show that language representation similarity reflect linguistic
similarity from multiple perspectives, including language family, geographical
sprachbund, lexicostatistics and syntax. Then we cluster all the target
languages into multiple groups and name each group as a representation
sprachbund. Thus, languages in the same representation sprachbund are supposed
to boost each other in both pre-training and fine-tuning as they share rich
linguistic similarity. We pre-train one multilingual model for each
representation sprachbund. Experiments are conducted on cross-lingual
benchmarks and significant improvements are achieved compared to strong
baselines.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz81KxDAUBeBsXMjoA7gyL9B689NJupTqqDCDot2Xm8yNBmpa0nbQt3ccXR0OHA58jF0JKLWtKrjB_BUPpRRQlwDSiHPW3MXJDwfKMb3zVxozTZRmnOOQ-NuY0X-4Je35Zsh8t_Rz7I-7BXv-kqloM8Z07BfsLGA_0eV_rli7uW-bx2L7_PDU3G4LXBtRqEqFWmmSeyu8RE1VQEMgg0NEHZwUwjlFtTVgHGlADXYdgsaapPVeqhW7_rs9Kboxx0_M392vpjtp1A-aS0YM</recordid><startdate>20210901</startdate><enddate>20210901</enddate><creator>Fan, Yimin</creator><creator>Liang, Yaobo</creator><creator>Muzio, Alexandre</creator><creator>Hassan, Hany</creator><creator>Li, Houqiang</creator><creator>Zhou, Ming</creator><creator>Duan, Nan</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20210901</creationdate><title>Discovering Representation Sprachbund For Multilingual Pre-Training</title><author>Fan, Yimin ; Liang, Yaobo ; Muzio, Alexandre ; Hassan, Hany ; Li, Houqiang ; Zhou, Ming ; Duan, Nan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-353f934e2d81c2a4e5fa7e02fbaaa4fb211bb3e98707be40a4086ff4a9e28cc23</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Fan, Yimin</creatorcontrib><creatorcontrib>Liang, Yaobo</creatorcontrib><creatorcontrib>Muzio, Alexandre</creatorcontrib><creatorcontrib>Hassan, Hany</creatorcontrib><creatorcontrib>Li, Houqiang</creatorcontrib><creatorcontrib>Zhou, Ming</creatorcontrib><creatorcontrib>Duan, Nan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Fan, Yimin</au><au>Liang, Yaobo</au><au>Muzio, Alexandre</au><au>Hassan, Hany</au><au>Li, Houqiang</au><au>Zhou, Ming</au><au>Duan, Nan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Discovering Representation Sprachbund For Multilingual Pre-Training</atitle><date>2021-09-01</date><risdate>2021</risdate><abstract>Multilingual pre-trained models have demonstrated their effectiveness in many
multilingual NLP tasks and enabled zero-shot or few-shot transfer from
high-resource languages to low resource ones. However, due to significant
typological differences and contradictions between some languages, such models
usually perform poorly on many languages and cross-lingual settings, which
shows the difficulty of learning a single model to handle massive diverse
languages well at the same time. To alleviate this issue, we present a new
multilingual pre-training pipeline. We propose to generate language
representation from multilingual pre-trained models and conduct linguistic
analysis to show that language representation similarity reflect linguistic
similarity from multiple perspectives, including language family, geographical
sprachbund, lexicostatistics and syntax. Then we cluster all the target
languages into multiple groups and name each group as a representation
sprachbund. Thus, languages in the same representation sprachbund are supposed
to boost each other in both pre-training and fine-tuning as they share rich
linguistic similarity. We pre-train one multilingual model for each
representation sprachbund. Experiments are conducted on cross-lingual
benchmarks and significant improvements are achieved compared to strong
baselines.</abstract><doi>10.48550/arxiv.2109.00271</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2109.00271 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2109_00271 |
source | arXiv.org |
subjects | Computer Science - Computation and Language |
title | Discovering Representation Sprachbund For Multilingual Pre-Training |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-22T11%3A07%3A30IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Discovering%20Representation%20Sprachbund%20For%20Multilingual%20Pre-Training&rft.au=Fan,%20Yimin&rft.date=2021-09-01&rft_id=info:doi/10.48550/arxiv.2109.00271&rft_dat=%3Carxiv_GOX%3E2109_00271%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |