Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models

Large language models (LLMs) can achieve highly effective performance on various reasoning tasks by incorporating step-by-step chain-of-thought (CoT) prompting as demonstrations. However, the reasoning chains of demonstrations generated by LLMs are prone to errors, which can subsequently lead to inc...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Sun, Jiashuo, Luo, Yi, Gong, Yeyun, Lin, Chen, Shen, Yelong, Guo, Jian, Duan, Nan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Sun, Jiashuo
Luo, Yi
Gong, Yeyun
Lin, Chen
Shen, Yelong
Guo, Jian
Duan, Nan
description Large language models (LLMs) can achieve highly effective performance on various reasoning tasks by incorporating step-by-step chain-of-thought (CoT) prompting as demonstrations. However, the reasoning chains of demonstrations generated by LLMs are prone to errors, which can subsequently lead to incorrect reasoning during inference. Furthermore, inappropriate exemplars (overly simplistic or complex), can affect overall performance among varying levels of difficulty. We introduce Iter-CoT (Iterative bootstrapping in Chain-of-Thoughts Prompting), an iterative bootstrapping approach for selecting exemplars and generating reasoning chains. By utilizing iterative bootstrapping, our approach enables LLMs to autonomously rectify errors, resulting in more precise and comprehensive reasoning chains. Simultaneously, our approach selects challenging yet answerable questions accompanied by reasoning chains as exemplars with a moderate level of difficulty, which enhances the LLMs' generalizability across varying levels of difficulty. Experimental results indicate that Iter-CoT exhibits superiority, achieving competitive performance across three distinct reasoning tasks on ten datasets.
doi_str_mv 10.48550/arxiv.2304.11657
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2304_11657</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2304_11657</sourcerecordid><originalsourceid>FETCH-LOGICAL-a677-8bbc93d3cb022a3e106b27da33e5569d117c47b7e4038c393010bbc30a79b8423</originalsourceid><addsrcrecordid>eNotj8FOhDAYhHvxYFYfwJO8QLHlLxSOSlbdBKMHbh7I39KFJrstKd1V315YvcxMMplJPkLuOEtFmefsAcO3PacZMJFyXuTymnxu3YhOWzck9YjWUb-n7ehPwxjn5CP44xTX7svGMdlFEzDas0mevI9zDDhNa2ld0mAYzKJuOOES3nxvDvMNudrjYTa3_74h7fO2rV9p8_6yqx8bioWUtFRKV9CDVizLEAxnhcpkjwAmz4uq51xqIZU0gkGpoQLG2TIBhrJSpchgQ-7_bi903RTsEcNPt1J2F0r4BTAMTVQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models</title><source>arXiv.org</source><creator>Sun, Jiashuo ; Luo, Yi ; Gong, Yeyun ; Lin, Chen ; Shen, Yelong ; Guo, Jian ; Duan, Nan</creator><creatorcontrib>Sun, Jiashuo ; Luo, Yi ; Gong, Yeyun ; Lin, Chen ; Shen, Yelong ; Guo, Jian ; Duan, Nan</creatorcontrib><description>Large language models (LLMs) can achieve highly effective performance on various reasoning tasks by incorporating step-by-step chain-of-thought (CoT) prompting as demonstrations. However, the reasoning chains of demonstrations generated by LLMs are prone to errors, which can subsequently lead to incorrect reasoning during inference. Furthermore, inappropriate exemplars (overly simplistic or complex), can affect overall performance among varying levels of difficulty. We introduce Iter-CoT (Iterative bootstrapping in Chain-of-Thoughts Prompting), an iterative bootstrapping approach for selecting exemplars and generating reasoning chains. By utilizing iterative bootstrapping, our approach enables LLMs to autonomously rectify errors, resulting in more precise and comprehensive reasoning chains. Simultaneously, our approach selects challenging yet answerable questions accompanied by reasoning chains as exemplars with a moderate level of difficulty, which enhances the LLMs' generalizability across varying levels of difficulty. Experimental results indicate that Iter-CoT exhibits superiority, achieving competitive performance across three distinct reasoning tasks on ten datasets.</description><identifier>DOI: 10.48550/arxiv.2304.11657</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computation and Language</subject><creationdate>2023-04</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2304.11657$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2304.11657$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Sun, Jiashuo</creatorcontrib><creatorcontrib>Luo, Yi</creatorcontrib><creatorcontrib>Gong, Yeyun</creatorcontrib><creatorcontrib>Lin, Chen</creatorcontrib><creatorcontrib>Shen, Yelong</creatorcontrib><creatorcontrib>Guo, Jian</creatorcontrib><creatorcontrib>Duan, Nan</creatorcontrib><title>Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models</title><description>Large language models (LLMs) can achieve highly effective performance on various reasoning tasks by incorporating step-by-step chain-of-thought (CoT) prompting as demonstrations. However, the reasoning chains of demonstrations generated by LLMs are prone to errors, which can subsequently lead to incorrect reasoning during inference. Furthermore, inappropriate exemplars (overly simplistic or complex), can affect overall performance among varying levels of difficulty. We introduce Iter-CoT (Iterative bootstrapping in Chain-of-Thoughts Prompting), an iterative bootstrapping approach for selecting exemplars and generating reasoning chains. By utilizing iterative bootstrapping, our approach enables LLMs to autonomously rectify errors, resulting in more precise and comprehensive reasoning chains. Simultaneously, our approach selects challenging yet answerable questions accompanied by reasoning chains as exemplars with a moderate level of difficulty, which enhances the LLMs' generalizability across varying levels of difficulty. Experimental results indicate that Iter-CoT exhibits superiority, achieving competitive performance across three distinct reasoning tasks on ten datasets.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj8FOhDAYhHvxYFYfwJO8QLHlLxSOSlbdBKMHbh7I39KFJrstKd1V315YvcxMMplJPkLuOEtFmefsAcO3PacZMJFyXuTymnxu3YhOWzck9YjWUb-n7ehPwxjn5CP44xTX7svGMdlFEzDas0mevI9zDDhNa2ld0mAYzKJuOOES3nxvDvMNudrjYTa3_74h7fO2rV9p8_6yqx8bioWUtFRKV9CDVizLEAxnhcpkjwAmz4uq51xqIZU0gkGpoQLG2TIBhrJSpchgQ-7_bi903RTsEcNPt1J2F0r4BTAMTVQ</recordid><startdate>20230423</startdate><enddate>20230423</enddate><creator>Sun, Jiashuo</creator><creator>Luo, Yi</creator><creator>Gong, Yeyun</creator><creator>Lin, Chen</creator><creator>Shen, Yelong</creator><creator>Guo, Jian</creator><creator>Duan, Nan</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230423</creationdate><title>Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models</title><author>Sun, Jiashuo ; Luo, Yi ; Gong, Yeyun ; Lin, Chen ; Shen, Yelong ; Guo, Jian ; Duan, Nan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a677-8bbc93d3cb022a3e106b27da33e5569d117c47b7e4038c393010bbc30a79b8423</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Sun, Jiashuo</creatorcontrib><creatorcontrib>Luo, Yi</creatorcontrib><creatorcontrib>Gong, Yeyun</creatorcontrib><creatorcontrib>Lin, Chen</creatorcontrib><creatorcontrib>Shen, Yelong</creatorcontrib><creatorcontrib>Guo, Jian</creatorcontrib><creatorcontrib>Duan, Nan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Sun, Jiashuo</au><au>Luo, Yi</au><au>Gong, Yeyun</au><au>Lin, Chen</au><au>Shen, Yelong</au><au>Guo, Jian</au><au>Duan, Nan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models</atitle><date>2023-04-23</date><risdate>2023</risdate><abstract>Large language models (LLMs) can achieve highly effective performance on various reasoning tasks by incorporating step-by-step chain-of-thought (CoT) prompting as demonstrations. However, the reasoning chains of demonstrations generated by LLMs are prone to errors, which can subsequently lead to incorrect reasoning during inference. Furthermore, inappropriate exemplars (overly simplistic or complex), can affect overall performance among varying levels of difficulty. We introduce Iter-CoT (Iterative bootstrapping in Chain-of-Thoughts Prompting), an iterative bootstrapping approach for selecting exemplars and generating reasoning chains. By utilizing iterative bootstrapping, our approach enables LLMs to autonomously rectify errors, resulting in more precise and comprehensive reasoning chains. Simultaneously, our approach selects challenging yet answerable questions accompanied by reasoning chains as exemplars with a moderate level of difficulty, which enhances the LLMs' generalizability across varying levels of difficulty. Experimental results indicate that Iter-CoT exhibits superiority, achieving competitive performance across three distinct reasoning tasks on ten datasets.</abstract><doi>10.48550/arxiv.2304.11657</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2304.11657
ispartof
issn
language eng
recordid cdi_arxiv_primary_2304_11657
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computation and Language
title Enhancing Chain-of-Thoughts Prompting with Iterative Bootstrapping in Large Language Models
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-11T01%3A40%3A27IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Enhancing%20Chain-of-Thoughts%20Prompting%20with%20Iterative%20Bootstrapping%20in%20Large%20Language%20Models&rft.au=Sun,%20Jiashuo&rft.date=2023-04-23&rft_id=info:doi/10.48550/arxiv.2304.11657&rft_dat=%3Carxiv_GOX%3E2304_11657%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true