Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence

Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM). Prior work assumes information in such knowledge sources is consistent with each other, paying little attention to how models blend inform...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Chen, Hung-Ting, Zhang, Michael J. Q, Choi, Eunsol
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Chen, Hung-Ting
Zhang, Michael J. Q
Choi, Eunsol
description Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM). Prior work assumes information in such knowledge sources is consistent with each other, paying little attention to how models blend information stored in their LM parameters with that from retrieved evidence documents. In this paper, we simulate knowledge conflicts (i.e., where parametric knowledge suggests one answer and different passages suggest different answers) and examine model behaviors. We find retrieval performance heavily impacts which sources models rely on, and current models mostly rely on non-parametric knowledge in their best-performing settings. We discover a troubling trend that contradictions among knowledge sources affect model confidence only marginally. To address this issue, we present a new calibration study, where models are discouraged from presenting any single answer when presented with multiple conflicting answer candidates in retrieved evidences.
doi_str_mv 10.48550/arxiv.2210.13701
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2210_13701</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2210_13701</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-523d2aa65d418dbe9f93f531ebbbb0bca7868ebb1b32e103102d633c13502ceb3</originalsourceid><addsrcrecordid>eNpNj8tOwzAQRb1hgQofwAr_QIrtqZOUHUTloRYhle4jP8bFkmtXTijl70kKQtzN6D400iHkirPprJaS3ah89IepEEPAoWL8nMS1N-90GdNnQLtF-pY-ssGO3mcft7RJu33A47--SdEFb_rulq7RqOB1Vv04fUkWQ0f7NOQuoOn_pmO7OHiL0eAFOXMqdHj5eydk87DYNE_F6vXxublbFaqseCEFWKFUKe2M11bj3M3BSeCoBzFtVFWX9WC4BoGcAWfClgCGg2TCoIYJuf55ewJu99nvVP5qR_D2BA7fKvxU4w</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence</title><source>arXiv.org</source><creator>Chen, Hung-Ting ; Zhang, Michael J. Q ; Choi, Eunsol</creator><creatorcontrib>Chen, Hung-Ting ; Zhang, Michael J. Q ; Choi, Eunsol</creatorcontrib><description>Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM). Prior work assumes information in such knowledge sources is consistent with each other, paying little attention to how models blend information stored in their LM parameters with that from retrieved evidence documents. In this paper, we simulate knowledge conflicts (i.e., where parametric knowledge suggests one answer and different passages suggest different answers) and examine model behaviors. We find retrieval performance heavily impacts which sources models rely on, and current models mostly rely on non-parametric knowledge in their best-performing settings. We discover a troubling trend that contradictions among knowledge sources affect model confidence only marginally. To address this issue, we present a new calibration study, where models are discouraged from presenting any single answer when presented with multiple conflicting answer candidates in retrieved evidences.</description><identifier>DOI: 10.48550/arxiv.2210.13701</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2022-10</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2210.13701$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2210.13701$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chen, Hung-Ting</creatorcontrib><creatorcontrib>Zhang, Michael J. Q</creatorcontrib><creatorcontrib>Choi, Eunsol</creatorcontrib><title>Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence</title><description>Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM). Prior work assumes information in such knowledge sources is consistent with each other, paying little attention to how models blend information stored in their LM parameters with that from retrieved evidence documents. In this paper, we simulate knowledge conflicts (i.e., where parametric knowledge suggests one answer and different passages suggest different answers) and examine model behaviors. We find retrieval performance heavily impacts which sources models rely on, and current models mostly rely on non-parametric knowledge in their best-performing settings. We discover a troubling trend that contradictions among knowledge sources affect model confidence only marginally. To address this issue, we present a new calibration study, where models are discouraged from presenting any single answer when presented with multiple conflicting answer candidates in retrieved evidences.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpNj8tOwzAQRb1hgQofwAr_QIrtqZOUHUTloRYhle4jP8bFkmtXTijl70kKQtzN6D400iHkirPprJaS3ah89IepEEPAoWL8nMS1N-90GdNnQLtF-pY-ssGO3mcft7RJu33A47--SdEFb_rulq7RqOB1Vv04fUkWQ0f7NOQuoOn_pmO7OHiL0eAFOXMqdHj5eydk87DYNE_F6vXxublbFaqseCEFWKFUKe2M11bj3M3BSeCoBzFtVFWX9WC4BoGcAWfClgCGg2TCoIYJuf55ewJu99nvVP5qR_D2BA7fKvxU4w</recordid><startdate>20221024</startdate><enddate>20221024</enddate><creator>Chen, Hung-Ting</creator><creator>Zhang, Michael J. Q</creator><creator>Choi, Eunsol</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20221024</creationdate><title>Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence</title><author>Chen, Hung-Ting ; Zhang, Michael J. Q ; Choi, Eunsol</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-523d2aa65d418dbe9f93f531ebbbb0bca7868ebb1b32e103102d633c13502ceb3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Chen, Hung-Ting</creatorcontrib><creatorcontrib>Zhang, Michael J. Q</creatorcontrib><creatorcontrib>Choi, Eunsol</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chen, Hung-Ting</au><au>Zhang, Michael J. Q</au><au>Choi, Eunsol</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence</atitle><date>2022-10-24</date><risdate>2022</risdate><abstract>Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM). Prior work assumes information in such knowledge sources is consistent with each other, paying little attention to how models blend information stored in their LM parameters with that from retrieved evidence documents. In this paper, we simulate knowledge conflicts (i.e., where parametric knowledge suggests one answer and different passages suggest different answers) and examine model behaviors. We find retrieval performance heavily impacts which sources models rely on, and current models mostly rely on non-parametric knowledge in their best-performing settings. We discover a troubling trend that contradictions among knowledge sources affect model confidence only marginally. To address this issue, we present a new calibration study, where models are discouraged from presenting any single answer when presented with multiple conflicting answer candidates in retrieved evidences.</abstract><doi>10.48550/arxiv.2210.13701</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2210.13701
ispartof
issn
language eng
recordid cdi_arxiv_primary_2210_13701
source arXiv.org
subjects Computer Science - Computation and Language
title Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T18%3A17%3A58IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Rich%20Knowledge%20Sources%20Bring%20Complex%20Knowledge%20Conflicts:%20Recalibrating%20Models%20to%20Reflect%20Conflicting%20Evidence&rft.au=Chen,%20Hung-Ting&rft.date=2022-10-24&rft_id=info:doi/10.48550/arxiv.2210.13701&rft_dat=%3Carxiv_GOX%3E2210_13701%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true