Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence
Question answering models can use rich knowledge sources -- up to one hundred retrieved passages and parametric knowledge in the large-scale language model (LM). Prior work assumes information in such knowledge sources is consistent with each other, paying little attention to how models blend inform...
Gespeichert in:
Hauptverfasser: | , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Chen, Hung-Ting Zhang, Michael J. Q Choi, Eunsol |
description | Question answering models can use rich knowledge sources -- up to one hundred
retrieved passages and parametric knowledge in the large-scale language model
(LM). Prior work assumes information in such knowledge sources is consistent
with each other, paying little attention to how models blend information stored
in their LM parameters with that from retrieved evidence documents. In this
paper, we simulate knowledge conflicts (i.e., where parametric knowledge
suggests one answer and different passages suggest different answers) and
examine model behaviors. We find retrieval performance heavily impacts which
sources models rely on, and current models mostly rely on non-parametric
knowledge in their best-performing settings. We discover a troubling trend that
contradictions among knowledge sources affect model confidence only marginally.
To address this issue, we present a new calibration study, where models are
discouraged from presenting any single answer when presented with multiple
conflicting answer candidates in retrieved evidences. |
doi_str_mv | 10.48550/arxiv.2210.13701 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2210_13701</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2210_13701</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-523d2aa65d418dbe9f93f531ebbbb0bca7868ebb1b32e103102d633c13502ceb3</originalsourceid><addsrcrecordid>eNpNj8tOwzAQRb1hgQofwAr_QIrtqZOUHUTloRYhle4jP8bFkmtXTijl70kKQtzN6D400iHkirPprJaS3ah89IepEEPAoWL8nMS1N-90GdNnQLtF-pY-ssGO3mcft7RJu33A47--SdEFb_rulq7RqOB1Vv04fUkWQ0f7NOQuoOn_pmO7OHiL0eAFOXMqdHj5eydk87DYNE_F6vXxublbFaqseCEFWKFUKe2M11bj3M3BSeCoBzFtVFWX9WC4BoGcAWfClgCGg2TCoIYJuf55ewJu99nvVP5qR_D2BA7fKvxU4w</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence</title><source>arXiv.org</source><creator>Chen, Hung-Ting ; Zhang, Michael J. Q ; Choi, Eunsol</creator><creatorcontrib>Chen, Hung-Ting ; Zhang, Michael J. Q ; Choi, Eunsol</creatorcontrib><description>Question answering models can use rich knowledge sources -- up to one hundred
retrieved passages and parametric knowledge in the large-scale language model
(LM). Prior work assumes information in such knowledge sources is consistent
with each other, paying little attention to how models blend information stored
in their LM parameters with that from retrieved evidence documents. In this
paper, we simulate knowledge conflicts (i.e., where parametric knowledge
suggests one answer and different passages suggest different answers) and
examine model behaviors. We find retrieval performance heavily impacts which
sources models rely on, and current models mostly rely on non-parametric
knowledge in their best-performing settings. We discover a troubling trend that
contradictions among knowledge sources affect model confidence only marginally.
To address this issue, we present a new calibration study, where models are
discouraged from presenting any single answer when presented with multiple
conflicting answer candidates in retrieved evidences.</description><identifier>DOI: 10.48550/arxiv.2210.13701</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2022-10</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2210.13701$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2210.13701$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Chen, Hung-Ting</creatorcontrib><creatorcontrib>Zhang, Michael J. Q</creatorcontrib><creatorcontrib>Choi, Eunsol</creatorcontrib><title>Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence</title><description>Question answering models can use rich knowledge sources -- up to one hundred
retrieved passages and parametric knowledge in the large-scale language model
(LM). Prior work assumes information in such knowledge sources is consistent
with each other, paying little attention to how models blend information stored
in their LM parameters with that from retrieved evidence documents. In this
paper, we simulate knowledge conflicts (i.e., where parametric knowledge
suggests one answer and different passages suggest different answers) and
examine model behaviors. We find retrieval performance heavily impacts which
sources models rely on, and current models mostly rely on non-parametric
knowledge in their best-performing settings. We discover a troubling trend that
contradictions among knowledge sources affect model confidence only marginally.
To address this issue, we present a new calibration study, where models are
discouraged from presenting any single answer when presented with multiple
conflicting answer candidates in retrieved evidences.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNpNj8tOwzAQRb1hgQofwAr_QIrtqZOUHUTloRYhle4jP8bFkmtXTijl70kKQtzN6D400iHkirPprJaS3ah89IepEEPAoWL8nMS1N-90GdNnQLtF-pY-ssGO3mcft7RJu33A47--SdEFb_rulq7RqOB1Vv04fUkWQ0f7NOQuoOn_pmO7OHiL0eAFOXMqdHj5eydk87DYNE_F6vXxublbFaqseCEFWKFUKe2M11bj3M3BSeCoBzFtVFWX9WC4BoGcAWfClgCGg2TCoIYJuf55ewJu99nvVP5qR_D2BA7fKvxU4w</recordid><startdate>20221024</startdate><enddate>20221024</enddate><creator>Chen, Hung-Ting</creator><creator>Zhang, Michael J. Q</creator><creator>Choi, Eunsol</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20221024</creationdate><title>Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence</title><author>Chen, Hung-Ting ; Zhang, Michael J. Q ; Choi, Eunsol</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-523d2aa65d418dbe9f93f531ebbbb0bca7868ebb1b32e103102d633c13502ceb3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Chen, Hung-Ting</creatorcontrib><creatorcontrib>Zhang, Michael J. Q</creatorcontrib><creatorcontrib>Choi, Eunsol</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Chen, Hung-Ting</au><au>Zhang, Michael J. Q</au><au>Choi, Eunsol</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence</atitle><date>2022-10-24</date><risdate>2022</risdate><abstract>Question answering models can use rich knowledge sources -- up to one hundred
retrieved passages and parametric knowledge in the large-scale language model
(LM). Prior work assumes information in such knowledge sources is consistent
with each other, paying little attention to how models blend information stored
in their LM parameters with that from retrieved evidence documents. In this
paper, we simulate knowledge conflicts (i.e., where parametric knowledge
suggests one answer and different passages suggest different answers) and
examine model behaviors. We find retrieval performance heavily impacts which
sources models rely on, and current models mostly rely on non-parametric
knowledge in their best-performing settings. We discover a troubling trend that
contradictions among knowledge sources affect model confidence only marginally.
To address this issue, we present a new calibration study, where models are
discouraged from presenting any single answer when presented with multiple
conflicting answer candidates in retrieved evidences.</abstract><doi>10.48550/arxiv.2210.13701</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2210.13701 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2210_13701 |
source | arXiv.org |
subjects | Computer Science - Computation and Language |
title | Rich Knowledge Sources Bring Complex Knowledge Conflicts: Recalibrating Models to Reflect Conflicting Evidence |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-18T18%3A17%3A58IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Rich%20Knowledge%20Sources%20Bring%20Complex%20Knowledge%20Conflicts:%20Recalibrating%20Models%20to%20Reflect%20Conflicting%20Evidence&rft.au=Chen,%20Hung-Ting&rft.date=2022-10-24&rft_id=info:doi/10.48550/arxiv.2210.13701&rft_dat=%3Carxiv_GOX%3E2210_13701%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |