FLARE: Faithful Logic-Aided Reasoning and Exploration

Modern Question Answering (QA) and Reasoning approaches based on Large Language Models (LLMs) commonly use prompting techniques, such as Chain-of-Thought (CoT), assuming the resulting generation will have a more granular exploration and reasoning over the question space and scope. However, such meth...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2024-10
Hauptverfasser: Arakelyan, Erik, Minervini, Pasquale, Verga, Pat, Lewis, Patrick, Augenstein, Isabelle
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Arakelyan, Erik
Minervini, Pasquale
Verga, Pat
Lewis, Patrick
Augenstein, Isabelle
description Modern Question Answering (QA) and Reasoning approaches based on Large Language Models (LLMs) commonly use prompting techniques, such as Chain-of-Thought (CoT), assuming the resulting generation will have a more granular exploration and reasoning over the question space and scope. However, such methods struggle with generating outputs that are faithful to the intermediate chain of reasoning produced by the model. On the other end of the spectrum, neuro-symbolic methods such as Faithful CoT (F-CoT) propose to combine LLMs with external symbolic solvers. While such approaches boast a high degree of faithfulness, they usually require a model trained for code generation and struggle with tasks that are ambiguous or hard to formalise strictly. We introduce \(\textbf{F}\)aithful \(\textbf{L}\)ogic-\(\textbf{A}\)ided \(\textbf{R}\)easoning and \(\textbf{E}\)xploration (\(\textbf{FLARE}\)), a novel interpretable approach for traversing the problem space using task decompositions. We use the LLM to plan a solution, soft-formalise the query into facts and predicates using a logic programming code and simulate that code execution using an exhaustive multi-hop search over the defined space. Our method allows us to compute the faithfulness of the reasoning process w.r.t. the generated code and analyse the steps of the multi-hop search without relying on external solvers. Our methods achieve SOTA results on \(\mathbf{7}\) out of \(\mathbf{9}\) diverse reasoning benchmarks. We also show that model faithfulness positively correlates with overall performance and further demonstrate that \(\textbf{FLARE}\) allows pinpointing the decisive factors sufficient for and leading to the correct answer with optimal reasoning during the multi-hop search.
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_3119295081</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>3119295081</sourcerecordid><originalsourceid>FETCH-proquest_journals_31192950813</originalsourceid><addsrcrecordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mQwdfNxDHK1UnBLzCzJSCvNUfDJT89M1nXMTElNUQhKTSzOz8vMS1dIzEtRcK0oyMkvSizJzM_jYWBNS8wpTuWF0twMym6uIc4eugVF-YWlqcUl8Vn5pUV5QKl4Y0NDSyNLUwMLQ2PiVAEA908z3w</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3119295081</pqid></control><display><type>article</type><title>FLARE: Faithful Logic-Aided Reasoning and Exploration</title><source>Free E- Journals</source><creator>Arakelyan, Erik ; Minervini, Pasquale ; Verga, Pat ; Lewis, Patrick ; Augenstein, Isabelle</creator><creatorcontrib>Arakelyan, Erik ; Minervini, Pasquale ; Verga, Pat ; Lewis, Patrick ; Augenstein, Isabelle</creatorcontrib><description>Modern Question Answering (QA) and Reasoning approaches based on Large Language Models (LLMs) commonly use prompting techniques, such as Chain-of-Thought (CoT), assuming the resulting generation will have a more granular exploration and reasoning over the question space and scope. However, such methods struggle with generating outputs that are faithful to the intermediate chain of reasoning produced by the model. On the other end of the spectrum, neuro-symbolic methods such as Faithful CoT (F-CoT) propose to combine LLMs with external symbolic solvers. While such approaches boast a high degree of faithfulness, they usually require a model trained for code generation and struggle with tasks that are ambiguous or hard to formalise strictly. We introduce \(\textbf{F}\)aithful \(\textbf{L}\)ogic-\(\textbf{A}\)ided \(\textbf{R}\)easoning and \(\textbf{E}\)xploration (\(\textbf{FLARE}\)), a novel interpretable approach for traversing the problem space using task decompositions. We use the LLM to plan a solution, soft-formalise the query into facts and predicates using a logic programming code and simulate that code execution using an exhaustive multi-hop search over the defined space. Our method allows us to compute the faithfulness of the reasoning process w.r.t. the generated code and analyse the steps of the multi-hop search without relying on external solvers. Our methods achieve SOTA results on \(\mathbf{7}\) out of \(\mathbf{9}\) diverse reasoning benchmarks. We also show that model faithfulness positively correlates with overall performance and further demonstrate that \(\textbf{FLARE}\) allows pinpointing the decisive factors sufficient for and leading to the correct answer with optimal reasoning during the multi-hop search.</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Large language models ; Logic programming ; Questions ; Reasoning ; Searching ; Solvers</subject><ispartof>arXiv.org, 2024-10</ispartof><rights>2024. This work is published under http://arxiv.org/licenses/nonexclusive-distrib/1.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>777,781</link.rule.ids></links><search><creatorcontrib>Arakelyan, Erik</creatorcontrib><creatorcontrib>Minervini, Pasquale</creatorcontrib><creatorcontrib>Verga, Pat</creatorcontrib><creatorcontrib>Lewis, Patrick</creatorcontrib><creatorcontrib>Augenstein, Isabelle</creatorcontrib><title>FLARE: Faithful Logic-Aided Reasoning and Exploration</title><title>arXiv.org</title><description>Modern Question Answering (QA) and Reasoning approaches based on Large Language Models (LLMs) commonly use prompting techniques, such as Chain-of-Thought (CoT), assuming the resulting generation will have a more granular exploration and reasoning over the question space and scope. However, such methods struggle with generating outputs that are faithful to the intermediate chain of reasoning produced by the model. On the other end of the spectrum, neuro-symbolic methods such as Faithful CoT (F-CoT) propose to combine LLMs with external symbolic solvers. While such approaches boast a high degree of faithfulness, they usually require a model trained for code generation and struggle with tasks that are ambiguous or hard to formalise strictly. We introduce \(\textbf{F}\)aithful \(\textbf{L}\)ogic-\(\textbf{A}\)ided \(\textbf{R}\)easoning and \(\textbf{E}\)xploration (\(\textbf{FLARE}\)), a novel interpretable approach for traversing the problem space using task decompositions. We use the LLM to plan a solution, soft-formalise the query into facts and predicates using a logic programming code and simulate that code execution using an exhaustive multi-hop search over the defined space. Our method allows us to compute the faithfulness of the reasoning process w.r.t. the generated code and analyse the steps of the multi-hop search without relying on external solvers. Our methods achieve SOTA results on \(\mathbf{7}\) out of \(\mathbf{9}\) diverse reasoning benchmarks. We also show that model faithfulness positively correlates with overall performance and further demonstrate that \(\textbf{FLARE}\) allows pinpointing the decisive factors sufficient for and leading to the correct answer with optimal reasoning during the multi-hop search.</description><subject>Large language models</subject><subject>Logic programming</subject><subject>Questions</subject><subject>Reasoning</subject><subject>Searching</subject><subject>Solvers</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><recordid>eNpjYuA0MjY21LUwMTLiYOAtLs4yMDAwMjM3MjU15mQwdfNxDHK1UnBLzCzJSCvNUfDJT89M1nXMTElNUQhKTSzOz8vMS1dIzEtRcK0oyMkvSizJzM_jYWBNS8wpTuWF0twMym6uIc4eugVF-YWlqcUl8Vn5pUV5QKl4Y0NDSyNLUwMLQ2PiVAEA908z3w</recordid><startdate>20241019</startdate><enddate>20241019</enddate><creator>Arakelyan, Erik</creator><creator>Minervini, Pasquale</creator><creator>Verga, Pat</creator><creator>Lewis, Patrick</creator><creator>Augenstein, Isabelle</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20241019</creationdate><title>FLARE: Faithful Logic-Aided Reasoning and Exploration</title><author>Arakelyan, Erik ; Minervini, Pasquale ; Verga, Pat ; Lewis, Patrick ; Augenstein, Isabelle</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_31192950813</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Large language models</topic><topic>Logic programming</topic><topic>Questions</topic><topic>Reasoning</topic><topic>Searching</topic><topic>Solvers</topic><toplevel>online_resources</toplevel><creatorcontrib>Arakelyan, Erik</creatorcontrib><creatorcontrib>Minervini, Pasquale</creatorcontrib><creatorcontrib>Verga, Pat</creatorcontrib><creatorcontrib>Lewis, Patrick</creatorcontrib><creatorcontrib>Augenstein, Isabelle</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Arakelyan, Erik</au><au>Minervini, Pasquale</au><au>Verga, Pat</au><au>Lewis, Patrick</au><au>Augenstein, Isabelle</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>FLARE: Faithful Logic-Aided Reasoning and Exploration</atitle><jtitle>arXiv.org</jtitle><date>2024-10-19</date><risdate>2024</risdate><eissn>2331-8422</eissn><abstract>Modern Question Answering (QA) and Reasoning approaches based on Large Language Models (LLMs) commonly use prompting techniques, such as Chain-of-Thought (CoT), assuming the resulting generation will have a more granular exploration and reasoning over the question space and scope. However, such methods struggle with generating outputs that are faithful to the intermediate chain of reasoning produced by the model. On the other end of the spectrum, neuro-symbolic methods such as Faithful CoT (F-CoT) propose to combine LLMs with external symbolic solvers. While such approaches boast a high degree of faithfulness, they usually require a model trained for code generation and struggle with tasks that are ambiguous or hard to formalise strictly. We introduce \(\textbf{F}\)aithful \(\textbf{L}\)ogic-\(\textbf{A}\)ided \(\textbf{R}\)easoning and \(\textbf{E}\)xploration (\(\textbf{FLARE}\)), a novel interpretable approach for traversing the problem space using task decompositions. We use the LLM to plan a solution, soft-formalise the query into facts and predicates using a logic programming code and simulate that code execution using an exhaustive multi-hop search over the defined space. Our method allows us to compute the faithfulness of the reasoning process w.r.t. the generated code and analyse the steps of the multi-hop search without relying on external solvers. Our methods achieve SOTA results on \(\mathbf{7}\) out of \(\mathbf{9}\) diverse reasoning benchmarks. We also show that model faithfulness positively correlates with overall performance and further demonstrate that \(\textbf{FLARE}\) allows pinpointing the decisive factors sufficient for and leading to the correct answer with optimal reasoning during the multi-hop search.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2024-10
issn 2331-8422
language eng
recordid cdi_proquest_journals_3119295081
source Free E- Journals
subjects Large language models
Logic programming
Questions
Reasoning
Searching
Solvers
title FLARE: Faithful Logic-Aided Reasoning and Exploration
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-20T04%3A00%3A32IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=FLARE:%20Faithful%20Logic-Aided%20Reasoning%20and%20Exploration&rft.jtitle=arXiv.org&rft.au=Arakelyan,%20Erik&rft.date=2024-10-19&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E3119295081%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3119295081&rft_id=info:pmid/&rfr_iscdi=true