Bootstrapping Multilingual Semantic Parsers using Large Language Models

Despite cross-lingual generalization demonstrated by pre-trained multilingual models, the translate-train paradigm of transferring English datasets across multiple languages remains to be a key mechanism for training task-specific multilingual models. However, for many low-resource languages, the av...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Awasthi, Abhijeet, Gupta, Nitish, Samanta, Bidisha, Dave, Shachi, Sarawagi, Sunita, Talukdar, Partha
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Awasthi, Abhijeet
Gupta, Nitish
Samanta, Bidisha
Dave, Shachi
Sarawagi, Sunita
Talukdar, Partha
description Despite cross-lingual generalization demonstrated by pre-trained multilingual models, the translate-train paradigm of transferring English datasets across multiple languages remains to be a key mechanism for training task-specific multilingual models. However, for many low-resource languages, the availability of a reliable translation service entails significant amounts of costly human-annotated translation pairs. Further, translation services may continue to be brittle due to domain mismatch between task-specific input text and general-purpose text used for training translation models. For multilingual semantic parsing, we demonstrate the effectiveness and flexibility offered by large language models (LLMs) for translating English datasets into several languages via few-shot prompting. Through extensive comparisons on two public datasets, MTOP and MASSIVE, spanning 50 languages and several domains, we show that our method of translating data using LLMs outperforms a strong translate-train baseline on 41 out of 50 languages. We study the key design choices that enable more effective multilingual data translation via prompted LLMs.
doi_str_mv 10.48550/arxiv.2210.07313
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2210_07313</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2210_07313</sourcerecordid><originalsourceid>FETCH-LOGICAL-a673-ec7202688dc83f75bbeab79617cf763387d70c6b829a06c693a1079d863a4e763</originalsourceid><addsrcrecordid>eNotj0FuwjAQRb3pAlEOwKq5QMDxEI-zbBEFpKBWKvto4jjIkiGRnVT09nVoN3--Rk9feowtM77aqDzna_J3-70SIj44QgYztn_ruiEMnvre3i7JaXSDdbGN5JIvc6XbYHXyST4YH5IxTExJ_mJiTlAsp64xLjyzp5ZcMIv_O2fn9915e0jLj_1x-1qmJBFSo1FwIZVqtIIW87o2VGMhM9QtSgCFDXItayUK4lLLAijjWDRKAm1MJObs5W_2YVL13l7J_1STUfUwgl-VM0Y2</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Bootstrapping Multilingual Semantic Parsers using Large Language Models</title><source>arXiv.org</source><creator>Awasthi, Abhijeet ; Gupta, Nitish ; Samanta, Bidisha ; Dave, Shachi ; Sarawagi, Sunita ; Talukdar, Partha</creator><creatorcontrib>Awasthi, Abhijeet ; Gupta, Nitish ; Samanta, Bidisha ; Dave, Shachi ; Sarawagi, Sunita ; Talukdar, Partha</creatorcontrib><description>Despite cross-lingual generalization demonstrated by pre-trained multilingual models, the translate-train paradigm of transferring English datasets across multiple languages remains to be a key mechanism for training task-specific multilingual models. However, for many low-resource languages, the availability of a reliable translation service entails significant amounts of costly human-annotated translation pairs. Further, translation services may continue to be brittle due to domain mismatch between task-specific input text and general-purpose text used for training translation models. For multilingual semantic parsing, we demonstrate the effectiveness and flexibility offered by large language models (LLMs) for translating English datasets into several languages via few-shot prompting. Through extensive comparisons on two public datasets, MTOP and MASSIVE, spanning 50 languages and several domains, we show that our method of translating data using LLMs outperforms a strong translate-train baseline on 41 out of 50 languages. We study the key design choices that enable more effective multilingual data translation via prompted LLMs.</description><identifier>DOI: 10.48550/arxiv.2210.07313</identifier><language>eng</language><subject>Computer Science - Computation and Language ; Computer Science - Learning</subject><creationdate>2022-10</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2210.07313$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2210.07313$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Awasthi, Abhijeet</creatorcontrib><creatorcontrib>Gupta, Nitish</creatorcontrib><creatorcontrib>Samanta, Bidisha</creatorcontrib><creatorcontrib>Dave, Shachi</creatorcontrib><creatorcontrib>Sarawagi, Sunita</creatorcontrib><creatorcontrib>Talukdar, Partha</creatorcontrib><title>Bootstrapping Multilingual Semantic Parsers using Large Language Models</title><description>Despite cross-lingual generalization demonstrated by pre-trained multilingual models, the translate-train paradigm of transferring English datasets across multiple languages remains to be a key mechanism for training task-specific multilingual models. However, for many low-resource languages, the availability of a reliable translation service entails significant amounts of costly human-annotated translation pairs. Further, translation services may continue to be brittle due to domain mismatch between task-specific input text and general-purpose text used for training translation models. For multilingual semantic parsing, we demonstrate the effectiveness and flexibility offered by large language models (LLMs) for translating English datasets into several languages via few-shot prompting. Through extensive comparisons on two public datasets, MTOP and MASSIVE, spanning 50 languages and several domains, we show that our method of translating data using LLMs outperforms a strong translate-train baseline on 41 out of 50 languages. We study the key design choices that enable more effective multilingual data translation via prompted LLMs.</description><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj0FuwjAQRb3pAlEOwKq5QMDxEI-zbBEFpKBWKvto4jjIkiGRnVT09nVoN3--Rk9feowtM77aqDzna_J3-70SIj44QgYztn_ruiEMnvre3i7JaXSDdbGN5JIvc6XbYHXyST4YH5IxTExJ_mJiTlAsp64xLjyzp5ZcMIv_O2fn9915e0jLj_1x-1qmJBFSo1FwIZVqtIIW87o2VGMhM9QtSgCFDXItayUK4lLLAijjWDRKAm1MJObs5W_2YVL13l7J_1STUfUwgl-VM0Y2</recordid><startdate>20221013</startdate><enddate>20221013</enddate><creator>Awasthi, Abhijeet</creator><creator>Gupta, Nitish</creator><creator>Samanta, Bidisha</creator><creator>Dave, Shachi</creator><creator>Sarawagi, Sunita</creator><creator>Talukdar, Partha</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20221013</creationdate><title>Bootstrapping Multilingual Semantic Parsers using Large Language Models</title><author>Awasthi, Abhijeet ; Gupta, Nitish ; Samanta, Bidisha ; Dave, Shachi ; Sarawagi, Sunita ; Talukdar, Partha</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a673-ec7202688dc83f75bbeab79617cf763387d70c6b829a06c693a1079d863a4e763</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Awasthi, Abhijeet</creatorcontrib><creatorcontrib>Gupta, Nitish</creatorcontrib><creatorcontrib>Samanta, Bidisha</creatorcontrib><creatorcontrib>Dave, Shachi</creatorcontrib><creatorcontrib>Sarawagi, Sunita</creatorcontrib><creatorcontrib>Talukdar, Partha</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Awasthi, Abhijeet</au><au>Gupta, Nitish</au><au>Samanta, Bidisha</au><au>Dave, Shachi</au><au>Sarawagi, Sunita</au><au>Talukdar, Partha</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Bootstrapping Multilingual Semantic Parsers using Large Language Models</atitle><date>2022-10-13</date><risdate>2022</risdate><abstract>Despite cross-lingual generalization demonstrated by pre-trained multilingual models, the translate-train paradigm of transferring English datasets across multiple languages remains to be a key mechanism for training task-specific multilingual models. However, for many low-resource languages, the availability of a reliable translation service entails significant amounts of costly human-annotated translation pairs. Further, translation services may continue to be brittle due to domain mismatch between task-specific input text and general-purpose text used for training translation models. For multilingual semantic parsing, we demonstrate the effectiveness and flexibility offered by large language models (LLMs) for translating English datasets into several languages via few-shot prompting. Through extensive comparisons on two public datasets, MTOP and MASSIVE, spanning 50 languages and several domains, we show that our method of translating data using LLMs outperforms a strong translate-train baseline on 41 out of 50 languages. We study the key design choices that enable more effective multilingual data translation via prompted LLMs.</abstract><doi>10.48550/arxiv.2210.07313</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2210.07313
ispartof
issn
language eng
recordid cdi_arxiv_primary_2210_07313
source arXiv.org
subjects Computer Science - Computation and Language
Computer Science - Learning
title Bootstrapping Multilingual Semantic Parsers using Large Language Models
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-06T14%3A09%3A11IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Bootstrapping%20Multilingual%20Semantic%20Parsers%20using%20Large%20Language%20Models&rft.au=Awasthi,%20Abhijeet&rft.date=2022-10-13&rft_id=info:doi/10.48550/arxiv.2210.07313&rft_dat=%3Carxiv_GOX%3E2210_07313%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true