Generate-then-Ground in Retrieval-Augmented Generation for Multi-hop Question Answering
Multi-Hop Question Answering (MHQA) tasks present a significant challenge for large language models (LLMs) due to the intensive knowledge required. Current solutions, like Retrieval-Augmented Generation, typically retrieve potential documents from an external corpus to read an answer. However, the p...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Shi, Zhengliang Sun, Weiwei Gao, Shen Ren, Pengjie Chen, Zhumin Ren, Zhaochun |
description | Multi-Hop Question Answering (MHQA) tasks present a significant challenge for
large language models (LLMs) due to the intensive knowledge required. Current
solutions, like Retrieval-Augmented Generation, typically retrieve potential
documents from an external corpus to read an answer. However, the performance
of this retrieve-then-read paradigm is constrained by the retriever and the
inevitable noise in the retrieved documents. To mitigate these challenges, we
introduce a novel generate-then-ground (GenGround) framework, synergizing the
parametric knowledge of LLMs and external documents to solve a multi-hop
question. GenGround empowers LLMs to alternate two phases until the final
answer is derived: (1) formulate a simpler, single-hop question and directly
generate the answer; (2) ground the question-answer pair in retrieved
documents, amending any wrong predictions in the answer. We also propose an
instructional grounding distillation method to generalize our method into
smaller models. Extensive experiments conducted on four datasets illustrate the
superiority of our method. |
doi_str_mv | 10.48550/arxiv.2406.14891 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2406_14891</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2406_14891</sourcerecordid><originalsourceid>FETCH-LOGICAL-a671-12320295fc97d1c232371d4e29aff0e70ad4afe836e9f9cade0f71d48427d0433</originalsourceid><addsrcrecordid>eNotj81KxDAUhbNxIaMP4Mq8QGr-2jTLMmgVRkQZcFkuzc1MoJMOaTrq2-vUWR3O4ePAR8id4IWuy5I_QPoOp0JqXhVC11Zck88WIybIyPIeI2vTOEdHQ6QfmFPAEwysmXcHjBkdvbBhjNSPib7OQw5sPx7p-4zTMjdx-sIU4u6GXHkYJry95Ipsnx6362e2eWtf1s2GQWUEE1JJLm3pe2uc6P-aMsJplBa852g4OA0ea1Wh9bYHh9yfgVpL47hWakXu_28Xs-6YwgHST3c27BZD9QviIEyI</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Generate-then-Ground in Retrieval-Augmented Generation for Multi-hop Question Answering</title><source>arXiv.org</source><creator>Shi, Zhengliang ; Sun, Weiwei ; Gao, Shen ; Ren, Pengjie ; Chen, Zhumin ; Ren, Zhaochun</creator><creatorcontrib>Shi, Zhengliang ; Sun, Weiwei ; Gao, Shen ; Ren, Pengjie ; Chen, Zhumin ; Ren, Zhaochun</creatorcontrib><description>Multi-Hop Question Answering (MHQA) tasks present a significant challenge for
large language models (LLMs) due to the intensive knowledge required. Current
solutions, like Retrieval-Augmented Generation, typically retrieve potential
documents from an external corpus to read an answer. However, the performance
of this retrieve-then-read paradigm is constrained by the retriever and the
inevitable noise in the retrieved documents. To mitigate these challenges, we
introduce a novel generate-then-ground (GenGround) framework, synergizing the
parametric knowledge of LLMs and external documents to solve a multi-hop
question. GenGround empowers LLMs to alternate two phases until the final
answer is derived: (1) formulate a simpler, single-hop question and directly
generate the answer; (2) ground the question-answer pair in retrieved
documents, amending any wrong predictions in the answer. We also propose an
instructional grounding distillation method to generalize our method into
smaller models. Extensive experiments conducted on four datasets illustrate the
superiority of our method.</description><identifier>DOI: 10.48550/arxiv.2406.14891</identifier><language>eng</language><subject>Computer Science - Computation and Language ; Computer Science - Information Retrieval</subject><creationdate>2024-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,777,882</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2406.14891$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2406.14891$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Shi, Zhengliang</creatorcontrib><creatorcontrib>Sun, Weiwei</creatorcontrib><creatorcontrib>Gao, Shen</creatorcontrib><creatorcontrib>Ren, Pengjie</creatorcontrib><creatorcontrib>Chen, Zhumin</creatorcontrib><creatorcontrib>Ren, Zhaochun</creatorcontrib><title>Generate-then-Ground in Retrieval-Augmented Generation for Multi-hop Question Answering</title><description>Multi-Hop Question Answering (MHQA) tasks present a significant challenge for
large language models (LLMs) due to the intensive knowledge required. Current
solutions, like Retrieval-Augmented Generation, typically retrieve potential
documents from an external corpus to read an answer. However, the performance
of this retrieve-then-read paradigm is constrained by the retriever and the
inevitable noise in the retrieved documents. To mitigate these challenges, we
introduce a novel generate-then-ground (GenGround) framework, synergizing the
parametric knowledge of LLMs and external documents to solve a multi-hop
question. GenGround empowers LLMs to alternate two phases until the final
answer is derived: (1) formulate a simpler, single-hop question and directly
generate the answer; (2) ground the question-answer pair in retrieved
documents, amending any wrong predictions in the answer. We also propose an
instructional grounding distillation method to generalize our method into
smaller models. Extensive experiments conducted on four datasets illustrate the
superiority of our method.</description><subject>Computer Science - Computation and Language</subject><subject>Computer Science - Information Retrieval</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj81KxDAUhbNxIaMP4Mq8QGr-2jTLMmgVRkQZcFkuzc1MoJMOaTrq2-vUWR3O4ePAR8id4IWuy5I_QPoOp0JqXhVC11Zck88WIybIyPIeI2vTOEdHQ6QfmFPAEwysmXcHjBkdvbBhjNSPib7OQw5sPx7p-4zTMjdx-sIU4u6GXHkYJry95Ipsnx6362e2eWtf1s2GQWUEE1JJLm3pe2uc6P-aMsJplBa852g4OA0ea1Wh9bYHh9yfgVpL47hWakXu_28Xs-6YwgHST3c27BZD9QviIEyI</recordid><startdate>20240621</startdate><enddate>20240621</enddate><creator>Shi, Zhengliang</creator><creator>Sun, Weiwei</creator><creator>Gao, Shen</creator><creator>Ren, Pengjie</creator><creator>Chen, Zhumin</creator><creator>Ren, Zhaochun</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240621</creationdate><title>Generate-then-Ground in Retrieval-Augmented Generation for Multi-hop Question Answering</title><author>Shi, Zhengliang ; Sun, Weiwei ; Gao, Shen ; Ren, Pengjie ; Chen, Zhumin ; Ren, Zhaochun</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a671-12320295fc97d1c232371d4e29aff0e70ad4afe836e9f9cade0f71d48427d0433</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Computation and Language</topic><topic>Computer Science - Information Retrieval</topic><toplevel>online_resources</toplevel><creatorcontrib>Shi, Zhengliang</creatorcontrib><creatorcontrib>Sun, Weiwei</creatorcontrib><creatorcontrib>Gao, Shen</creatorcontrib><creatorcontrib>Ren, Pengjie</creatorcontrib><creatorcontrib>Chen, Zhumin</creatorcontrib><creatorcontrib>Ren, Zhaochun</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Shi, Zhengliang</au><au>Sun, Weiwei</au><au>Gao, Shen</au><au>Ren, Pengjie</au><au>Chen, Zhumin</au><au>Ren, Zhaochun</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Generate-then-Ground in Retrieval-Augmented Generation for Multi-hop Question Answering</atitle><date>2024-06-21</date><risdate>2024</risdate><abstract>Multi-Hop Question Answering (MHQA) tasks present a significant challenge for
large language models (LLMs) due to the intensive knowledge required. Current
solutions, like Retrieval-Augmented Generation, typically retrieve potential
documents from an external corpus to read an answer. However, the performance
of this retrieve-then-read paradigm is constrained by the retriever and the
inevitable noise in the retrieved documents. To mitigate these challenges, we
introduce a novel generate-then-ground (GenGround) framework, synergizing the
parametric knowledge of LLMs and external documents to solve a multi-hop
question. GenGround empowers LLMs to alternate two phases until the final
answer is derived: (1) formulate a simpler, single-hop question and directly
generate the answer; (2) ground the question-answer pair in retrieved
documents, amending any wrong predictions in the answer. We also propose an
instructional grounding distillation method to generalize our method into
smaller models. Extensive experiments conducted on four datasets illustrate the
superiority of our method.</abstract><doi>10.48550/arxiv.2406.14891</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2406.14891 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2406_14891 |
source | arXiv.org |
subjects | Computer Science - Computation and Language Computer Science - Information Retrieval |
title | Generate-then-Ground in Retrieval-Augmented Generation for Multi-hop Question Answering |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-20T09%3A02%3A28IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Generate-then-Ground%20in%20Retrieval-Augmented%20Generation%20for%20Multi-hop%20Question%20Answering&rft.au=Shi,%20Zhengliang&rft.date=2024-06-21&rft_id=info:doi/10.48550/arxiv.2406.14891&rft_dat=%3Carxiv_GOX%3E2406_14891%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |