Position Matters! Empirical Study of Order Effect in Knowledge-grounded Dialogue

With the power of large pretrained language models, various research works have integrated knowledge into dialogue systems. The traditional techniques treat knowledge as part of the input sequence for the dialogue system, prepending a set of knowledge statements in front of dialogue history. However...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Su, Hsuan, Kumar, Shachi H, Mazumder, Sahisnu, Chen, Wenda, Manuvinakurike, Ramesh, Okur, Eda, Sahay, Saurav, Nachman, Lama, Chen, Shang-Tse, Lee, Hung-yi
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Su, Hsuan
Kumar, Shachi H
Mazumder, Sahisnu
Chen, Wenda
Manuvinakurike, Ramesh
Okur, Eda
Sahay, Saurav
Nachman, Lama
Chen, Shang-Tse
Lee, Hung-yi
description With the power of large pretrained language models, various research works have integrated knowledge into dialogue systems. The traditional techniques treat knowledge as part of the input sequence for the dialogue system, prepending a set of knowledge statements in front of dialogue history. However, such a mechanism forces knowledge sets to be concatenated in an ordered manner, making models implicitly pay imbalanced attention to the sets during training. In this paper, we first investigate how the order of the knowledge set can influence autoregressive dialogue systems' responses. We conduct experiments on two commonly used dialogue datasets with two types of transformer-based models and find that models view the input knowledge unequally. To this end, we propose a simple and novel technique to alleviate the order effect by modifying the position embeddings of knowledge input in these models. With the proposed position embedding method, the experimental results show that each knowledge statement is uniformly considered to generate responses.
doi_str_mv 10.48550/arxiv.2302.05888
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2302_05888</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2302_05888</sourcerecordid><originalsourceid>FETCH-LOGICAL-a678-1c3d7e3f295860547f2811ededbde02aab75271bff4f2dacb8230517367ac2413</originalsourceid><addsrcrecordid>eNotz0tOwzAUhWFPGKDCAhhhFpDgRxybISrhoRa1Ep1HN_Z1ZCmNK8cBuntKYXRmv85HyA1nZWWUYveQvsNnKSQTJVPGmEuy3cYp5BBH-g45Y5ruaLM_hBQsDPQjz-5Io6eb5DDRxnu0mYaRrsb4NaDrsehTnEeHjj4FGGI_4xW58DBMeP2_C7J7bnbL12K9eXlbPq4LqLUpuJVOo_TiQZmaqUp7YTjHU6hzyARAp5XQvPO-8sKB7czptOJa1hqsqLhckNu_7JnUHlLYQzq2v7T2TJM_HyRJbA</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Position Matters! Empirical Study of Order Effect in Knowledge-grounded Dialogue</title><source>arXiv.org</source><creator>Su, Hsuan ; Kumar, Shachi H ; Mazumder, Sahisnu ; Chen, Wenda ; Manuvinakurike, Ramesh ; Okur, Eda ; Sahay, Saurav ; Nachman, Lama ; Chen, Shang-Tse ; Lee, Hung-yi</creator><creatorcontrib>Su, Hsuan ; Kumar, Shachi H ; Mazumder, Sahisnu ; Chen, Wenda ; Manuvinakurike, Ramesh ; Okur, Eda ; Sahay, Saurav ; Nachman, Lama ; Chen, Shang-Tse ; Lee, Hung-yi</creatorcontrib><description>With the power of large pretrained language models, various research works have integrated knowledge into dialogue systems. The traditional techniques treat knowledge as part of the input sequence for the dialogue system, prepending a set of knowledge statements in front of dialogue history. However, such a mechanism forces knowledge sets to be concatenated in an ordered manner, making models implicitly pay imbalanced attention to the sets during training. In this paper, we first investigate how the order of the knowledge set can influence autoregressive dialogue systems' responses. We conduct experiments on two commonly used dialogue datasets with two types of transformer-based models and find that models view the input knowledge unequally. To this end, we propose a simple and novel technique to alleviate the order effect by modifying the position embeddings of knowledge input in these models. With the proposed position embedding method, the experimental results show that each knowledge statement is uniformly considered to generate responses.</description><identifier>DOI: 10.48550/arxiv.2302.05888</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language ; Computer Science - Learning</subject><creationdate>2023-02</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2302.05888$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2302.05888$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Su, Hsuan</creatorcontrib><creatorcontrib>Kumar, Shachi H</creatorcontrib><creatorcontrib>Mazumder, Sahisnu</creatorcontrib><creatorcontrib>Chen, Wenda</creatorcontrib><creatorcontrib>Manuvinakurike, Ramesh</creatorcontrib><creatorcontrib>Okur, Eda</creatorcontrib><creatorcontrib>Sahay, Saurav</creatorcontrib><creatorcontrib>Nachman, Lama</creatorcontrib><creatorcontrib>Chen, Shang-Tse</creatorcontrib><creatorcontrib>Lee, Hung-yi</creatorcontrib><title>Position Matters! Empirical Study of Order Effect in Knowledge-grounded Dialogue</title><description>With the power of large pretrained language models, various research works have integrated knowledge into dialogue systems. The traditional techniques treat knowledge as part of the input sequence for the dialogue system, prepending a set of knowledge statements in front of dialogue history. However, such a mechanism forces knowledge sets to be concatenated in an ordered manner, making models implicitly pay imbalanced attention to the sets during training. In this paper, we first investigate how the order of the knowledge set can influence autoregressive dialogue systems' responses. We conduct experiments on two commonly used dialogue datasets with two types of transformer-based models and find that models view the input knowledge unequally. To this end, we propose a simple and novel technique to alleviate the order effect by modifying the position embeddings of knowledge input in these models. With the proposed position embedding method, the experimental results show that each knowledge statement is uniformly considered to generate responses.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz0tOwzAUhWFPGKDCAhhhFpDgRxybISrhoRa1Ep1HN_Z1ZCmNK8cBuntKYXRmv85HyA1nZWWUYveQvsNnKSQTJVPGmEuy3cYp5BBH-g45Y5ruaLM_hBQsDPQjz-5Io6eb5DDRxnu0mYaRrsb4NaDrsehTnEeHjj4FGGI_4xW58DBMeP2_C7J7bnbL12K9eXlbPq4LqLUpuJVOo_TiQZmaqUp7YTjHU6hzyARAp5XQvPO-8sKB7czptOJa1hqsqLhckNu_7JnUHlLYQzq2v7T2TJM_HyRJbA</recordid><startdate>20230212</startdate><enddate>20230212</enddate><creator>Su, Hsuan</creator><creator>Kumar, Shachi H</creator><creator>Mazumder, Sahisnu</creator><creator>Chen, Wenda</creator><creator>Manuvinakurike, Ramesh</creator><creator>Okur, Eda</creator><creator>Sahay, Saurav</creator><creator>Nachman, Lama</creator><creator>Chen, Shang-Tse</creator><creator>Lee, Hung-yi</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230212</creationdate><title>Position Matters! Empirical Study of Order Effect in Knowledge-grounded Dialogue</title><author>Su, Hsuan ; Kumar, Shachi H ; Mazumder, Sahisnu ; Chen, Wenda ; Manuvinakurike, Ramesh ; Okur, Eda ; Sahay, Saurav ; Nachman, Lama ; Chen, Shang-Tse ; Lee, Hung-yi</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a678-1c3d7e3f295860547f2811ededbde02aab75271bff4f2dacb8230517367ac2413</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Su, Hsuan</creatorcontrib><creatorcontrib>Kumar, Shachi H</creatorcontrib><creatorcontrib>Mazumder, Sahisnu</creatorcontrib><creatorcontrib>Chen, Wenda</creatorcontrib><creatorcontrib>Manuvinakurike, Ramesh</creatorcontrib><creatorcontrib>Okur, Eda</creatorcontrib><creatorcontrib>Sahay, Saurav</creatorcontrib><creatorcontrib>Nachman, Lama</creatorcontrib><creatorcontrib>Chen, Shang-Tse</creatorcontrib><creatorcontrib>Lee, Hung-yi</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Su, Hsuan</au><au>Kumar, Shachi H</au><au>Mazumder, Sahisnu</au><au>Chen, Wenda</au><au>Manuvinakurike, Ramesh</au><au>Okur, Eda</au><au>Sahay, Saurav</au><au>Nachman, Lama</au><au>Chen, Shang-Tse</au><au>Lee, Hung-yi</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Position Matters! Empirical Study of Order Effect in Knowledge-grounded Dialogue</atitle><date>2023-02-12</date><risdate>2023</risdate><abstract>With the power of large pretrained language models, various research works have integrated knowledge into dialogue systems. The traditional techniques treat knowledge as part of the input sequence for the dialogue system, prepending a set of knowledge statements in front of dialogue history. However, such a mechanism forces knowledge sets to be concatenated in an ordered manner, making models implicitly pay imbalanced attention to the sets during training. In this paper, we first investigate how the order of the knowledge set can influence autoregressive dialogue systems' responses. We conduct experiments on two commonly used dialogue datasets with two types of transformer-based models and find that models view the input knowledge unequally. To this end, we propose a simple and novel technique to alleviate the order effect by modifying the position embeddings of knowledge input in these models. With the proposed position embedding method, the experimental results show that each knowledge statement is uniformly considered to generate responses.</abstract><doi>10.48550/arxiv.2302.05888</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2302.05888
ispartof
issn
language eng
recordid cdi_arxiv_primary_2302_05888
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computation and Language
Computer Science - Learning
title Position Matters! Empirical Study of Order Effect in Knowledge-grounded Dialogue
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T17%3A59%3A26IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Position%20Matters!%20Empirical%20Study%20of%20Order%20Effect%20in%20Knowledge-grounded%20Dialogue&rft.au=Su,%20Hsuan&rft.date=2023-02-12&rft_id=info:doi/10.48550/arxiv.2302.05888&rft_dat=%3Carxiv_GOX%3E2302_05888%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true