Breakpoint Transformers for Modeling and Tracking Intermediate Beliefs

Can we teach natural language understanding models to track their beliefs through intermediate points in text? We propose a representation learning framework called breakpoint modeling that allows for learning of this type. Given any text encoder and data marked with intermediate states (breakpoints...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:arXiv.org 2022-11
Hauptverfasser: Richardson, Kyle, Tamari, Ronen, Sultan, Oren, Tsarfaty, Reut, Shahaf, Dafna, Sabharwal, Ashish
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title arXiv.org
container_volume
creator Richardson, Kyle
Tamari, Ronen
Sultan, Oren
Tsarfaty, Reut
Shahaf, Dafna
Sabharwal, Ashish
description Can we teach natural language understanding models to track their beliefs through intermediate points in text? We propose a representation learning framework called breakpoint modeling that allows for learning of this type. Given any text encoder and data marked with intermediate states (breakpoints) along with corresponding textual queries viewed as true/false propositions (i.e., the candidate beliefs of a model, consisting of information changing through time) our approach trains models in an efficient and end-to-end fashion to build intermediate representations that facilitate teaching and direct querying of beliefs at arbitrary points alongside solving other end tasks. To show the benefit of our approach, we experiment with a diverse set of NLU tasks including relational reasoning on CLUTRR and narrative understanding on bAbI. Using novel belief prediction tasks for both tasks, we show the benefit of our main breakpoint transformer, based on T5, over conventional representation learning approaches in terms of processing efficiency, prediction accuracy and prediction consistency, all with minimal to no effect on corresponding QA end tasks. To show the feasibility of incorporating our belief tracker into more complex reasoning pipelines, we also obtain SOTA performance on the three-tiered reasoning challenge for the TRIP benchmark (around 23-32% absolute improvement on Tasks 2-3).
format Article
fullrecord <record><control><sourceid>proquest</sourceid><recordid>TN_cdi_proquest_journals_2736928306</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2736928306</sourcerecordid><originalsourceid>FETCH-proquest_journals_27369283063</originalsourceid><addsrcrecordid>eNqNiksKwjAUAIMgWLR3CLguxBf7cVux6MJd9yU0r9KPSc1L728KHsDVMMxsWARSnpLiDLBjMdEghIAshzSVEatKh2qcbW88r50y1Fn3Rkc8kD-txqk3L66MXms7rvIwHsOje-WRl2HAjg5s26mJMP5xz47Vrb7ek9nZz4Lkm8EuzoTUQC6zCxRSZPK_6wvlVzuN</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2736928306</pqid></control><display><type>article</type><title>Breakpoint Transformers for Modeling and Tracking Intermediate Beliefs</title><source>Free E- Journals</source><creator>Richardson, Kyle ; Tamari, Ronen ; Sultan, Oren ; Tsarfaty, Reut ; Shahaf, Dafna ; Sabharwal, Ashish</creator><creatorcontrib>Richardson, Kyle ; Tamari, Ronen ; Sultan, Oren ; Tsarfaty, Reut ; Shahaf, Dafna ; Sabharwal, Ashish</creatorcontrib><description>Can we teach natural language understanding models to track their beliefs through intermediate points in text? We propose a representation learning framework called breakpoint modeling that allows for learning of this type. Given any text encoder and data marked with intermediate states (breakpoints) along with corresponding textual queries viewed as true/false propositions (i.e., the candidate beliefs of a model, consisting of information changing through time) our approach trains models in an efficient and end-to-end fashion to build intermediate representations that facilitate teaching and direct querying of beliefs at arbitrary points alongside solving other end tasks. To show the benefit of our approach, we experiment with a diverse set of NLU tasks including relational reasoning on CLUTRR and narrative understanding on bAbI. Using novel belief prediction tasks for both tasks, we show the benefit of our main breakpoint transformer, based on T5, over conventional representation learning approaches in terms of processing efficiency, prediction accuracy and prediction consistency, all with minimal to no effect on corresponding QA end tasks. To show the feasibility of incorporating our belief tracker into more complex reasoning pipelines, we also obtain SOTA performance on the three-tiered reasoning challenge for the TRIP benchmark (around 23-32% absolute improvement on Tasks 2-3).</description><identifier>EISSN: 2331-8422</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Coders ; Learning ; Modelling ; Reasoning ; Representations ; Transformers</subject><ispartof>arXiv.org, 2022-11</ispartof><rights>2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>776,780</link.rule.ids></links><search><creatorcontrib>Richardson, Kyle</creatorcontrib><creatorcontrib>Tamari, Ronen</creatorcontrib><creatorcontrib>Sultan, Oren</creatorcontrib><creatorcontrib>Tsarfaty, Reut</creatorcontrib><creatorcontrib>Shahaf, Dafna</creatorcontrib><creatorcontrib>Sabharwal, Ashish</creatorcontrib><title>Breakpoint Transformers for Modeling and Tracking Intermediate Beliefs</title><title>arXiv.org</title><description>Can we teach natural language understanding models to track their beliefs through intermediate points in text? We propose a representation learning framework called breakpoint modeling that allows for learning of this type. Given any text encoder and data marked with intermediate states (breakpoints) along with corresponding textual queries viewed as true/false propositions (i.e., the candidate beliefs of a model, consisting of information changing through time) our approach trains models in an efficient and end-to-end fashion to build intermediate representations that facilitate teaching and direct querying of beliefs at arbitrary points alongside solving other end tasks. To show the benefit of our approach, we experiment with a diverse set of NLU tasks including relational reasoning on CLUTRR and narrative understanding on bAbI. Using novel belief prediction tasks for both tasks, we show the benefit of our main breakpoint transformer, based on T5, over conventional representation learning approaches in terms of processing efficiency, prediction accuracy and prediction consistency, all with minimal to no effect on corresponding QA end tasks. To show the feasibility of incorporating our belief tracker into more complex reasoning pipelines, we also obtain SOTA performance on the three-tiered reasoning challenge for the TRIP benchmark (around 23-32% absolute improvement on Tasks 2-3).</description><subject>Coders</subject><subject>Learning</subject><subject>Modelling</subject><subject>Reasoning</subject><subject>Representations</subject><subject>Transformers</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>BENPR</sourceid><recordid>eNqNiksKwjAUAIMgWLR3CLguxBf7cVux6MJd9yU0r9KPSc1L728KHsDVMMxsWARSnpLiDLBjMdEghIAshzSVEatKh2qcbW88r50y1Fn3Rkc8kD-txqk3L66MXms7rvIwHsOje-WRl2HAjg5s26mJMP5xz47Vrb7ek9nZz4Lkm8EuzoTUQC6zCxRSZPK_6wvlVzuN</recordid><startdate>20221115</startdate><enddate>20221115</enddate><creator>Richardson, Kyle</creator><creator>Tamari, Ronen</creator><creator>Sultan, Oren</creator><creator>Tsarfaty, Reut</creator><creator>Shahaf, Dafna</creator><creator>Sabharwal, Ashish</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope></search><sort><creationdate>20221115</creationdate><title>Breakpoint Transformers for Modeling and Tracking Intermediate Beliefs</title><author>Richardson, Kyle ; Tamari, Ronen ; Sultan, Oren ; Tsarfaty, Reut ; Shahaf, Dafna ; Sabharwal, Ashish</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-proquest_journals_27369283063</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Coders</topic><topic>Learning</topic><topic>Modelling</topic><topic>Reasoning</topic><topic>Representations</topic><topic>Transformers</topic><toplevel>online_resources</toplevel><creatorcontrib>Richardson, Kyle</creatorcontrib><creatorcontrib>Tamari, Ronen</creatorcontrib><creatorcontrib>Sultan, Oren</creatorcontrib><creatorcontrib>Tsarfaty, Reut</creatorcontrib><creatorcontrib>Shahaf, Dafna</creatorcontrib><creatorcontrib>Sabharwal, Ashish</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science &amp; Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Richardson, Kyle</au><au>Tamari, Ronen</au><au>Sultan, Oren</au><au>Tsarfaty, Reut</au><au>Shahaf, Dafna</au><au>Sabharwal, Ashish</au><format>book</format><genre>document</genre><ristype>GEN</ristype><atitle>Breakpoint Transformers for Modeling and Tracking Intermediate Beliefs</atitle><jtitle>arXiv.org</jtitle><date>2022-11-15</date><risdate>2022</risdate><eissn>2331-8422</eissn><abstract>Can we teach natural language understanding models to track their beliefs through intermediate points in text? We propose a representation learning framework called breakpoint modeling that allows for learning of this type. Given any text encoder and data marked with intermediate states (breakpoints) along with corresponding textual queries viewed as true/false propositions (i.e., the candidate beliefs of a model, consisting of information changing through time) our approach trains models in an efficient and end-to-end fashion to build intermediate representations that facilitate teaching and direct querying of beliefs at arbitrary points alongside solving other end tasks. To show the benefit of our approach, we experiment with a diverse set of NLU tasks including relational reasoning on CLUTRR and narrative understanding on bAbI. Using novel belief prediction tasks for both tasks, we show the benefit of our main breakpoint transformer, based on T5, over conventional representation learning approaches in terms of processing efficiency, prediction accuracy and prediction consistency, all with minimal to no effect on corresponding QA end tasks. To show the feasibility of incorporating our belief tracker into more complex reasoning pipelines, we also obtain SOTA performance on the three-tiered reasoning challenge for the TRIP benchmark (around 23-32% absolute improvement on Tasks 2-3).</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier EISSN: 2331-8422
ispartof arXiv.org, 2022-11
issn 2331-8422
language eng
recordid cdi_proquest_journals_2736928306
source Free E- Journals
subjects Coders
Learning
Modelling
Reasoning
Representations
Transformers
title Breakpoint Transformers for Modeling and Tracking Intermediate Beliefs
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-12T22%3A03%3A55IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest&rft_val_fmt=info:ofi/fmt:kev:mtx:book&rft.genre=document&rft.atitle=Breakpoint%20Transformers%20for%20Modeling%20and%20Tracking%20Intermediate%20Beliefs&rft.jtitle=arXiv.org&rft.au=Richardson,%20Kyle&rft.date=2022-11-15&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest%3E2736928306%3C/proquest%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2736928306&rft_id=info:pmid/&rfr_iscdi=true