Are You Sure? Challenging LLMs Leads to Performance Drops in The FlipFlop Experiment

The interactive nature of Large Language Models (LLMs) theoretically allows models to refine and improve their answers, yet systematic analysis of the multi-turn behavior of LLMs remains limited. In this paper, we propose the FlipFlop experiment: in the first round of the conversation, an LLM comple...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Laban, Philippe, Murakhovs'ka, Lidiya, Xiong, Caiming, Wu, Chien-Sheng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Laban, Philippe
Murakhovs'ka, Lidiya
Xiong, Caiming
Wu, Chien-Sheng
description The interactive nature of Large Language Models (LLMs) theoretically allows models to refine and improve their answers, yet systematic analysis of the multi-turn behavior of LLMs remains limited. In this paper, we propose the FlipFlop experiment: in the first round of the conversation, an LLM completes a classification task. In a second round, the LLM is challenged with a follow-up phrase like "Are you sure?", offering an opportunity for the model to reflect on its initial answer, and decide whether to confirm or flip its answer. A systematic study of ten LLMs on seven classification tasks reveals that models flip their answers on average 46% of the time and that all models see a deterioration of accuracy between their first and final prediction, with an average drop of 17% (the FlipFlop effect). We conduct finetuning experiments on an open-source LLM and find that finetuning on synthetically created data can mitigate - reducing performance deterioration by 60% - but not resolve sycophantic behavior entirely. The FlipFlop experiment illustrates the universality of sycophantic behavior in LLMs and provides a robust framework to analyze model behavior and evaluate future models.
doi_str_mv 10.48550/arxiv.2311.08596
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2311_08596</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2311_08596</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-e11b29f80bebdb846d65bd4fe4e39138610f0a65a914e8f235cb5d7c61c6b5c63</originalsourceid><addsrcrecordid>eNotz7FOwzAUhWEvDKjwAEzcF0iw49h1JlSFBpCMQGqWTpGdXLeRXMdyWlTeHihMZ_uPPkLuGM1LJQR9MOk8fuYFZyynSlTymrSrhLCdTrA5JXyEem-8x7Abww60fptBoxlmOE7wgclN6WBCj_CUpjjDGKDdIzR-jI2fIqzPEdN4wHC8IVfO-Blv_3dB2mbd1i-Zfn9-rVc6M3IpM2TMFpVT1KIdrCrlIIUdSocl8opxJRl11EhhKlaicgUXvRXDspesl1b0ki_I_V_2wuriz7lJX90vr7vw-DfGWEoo</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Are You Sure? Challenging LLMs Leads to Performance Drops in The FlipFlop Experiment</title><source>arXiv.org</source><creator>Laban, Philippe ; Murakhovs'ka, Lidiya ; Xiong, Caiming ; Wu, Chien-Sheng</creator><creatorcontrib>Laban, Philippe ; Murakhovs'ka, Lidiya ; Xiong, Caiming ; Wu, Chien-Sheng</creatorcontrib><description>The interactive nature of Large Language Models (LLMs) theoretically allows models to refine and improve their answers, yet systematic analysis of the multi-turn behavior of LLMs remains limited. In this paper, we propose the FlipFlop experiment: in the first round of the conversation, an LLM completes a classification task. In a second round, the LLM is challenged with a follow-up phrase like "Are you sure?", offering an opportunity for the model to reflect on its initial answer, and decide whether to confirm or flip its answer. A systematic study of ten LLMs on seven classification tasks reveals that models flip their answers on average 46% of the time and that all models see a deterioration of accuracy between their first and final prediction, with an average drop of 17% (the FlipFlop effect). We conduct finetuning experiments on an open-source LLM and find that finetuning on synthetically created data can mitigate - reducing performance deterioration by 60% - but not resolve sycophantic behavior entirely. The FlipFlop experiment illustrates the universality of sycophantic behavior in LLMs and provides a robust framework to analyze model behavior and evaluate future models.</description><identifier>DOI: 10.48550/arxiv.2311.08596</identifier><language>eng</language><subject>Computer Science - Computation and Language</subject><creationdate>2023-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2311.08596$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2311.08596$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Laban, Philippe</creatorcontrib><creatorcontrib>Murakhovs'ka, Lidiya</creatorcontrib><creatorcontrib>Xiong, Caiming</creatorcontrib><creatorcontrib>Wu, Chien-Sheng</creatorcontrib><title>Are You Sure? Challenging LLMs Leads to Performance Drops in The FlipFlop Experiment</title><description>The interactive nature of Large Language Models (LLMs) theoretically allows models to refine and improve their answers, yet systematic analysis of the multi-turn behavior of LLMs remains limited. In this paper, we propose the FlipFlop experiment: in the first round of the conversation, an LLM completes a classification task. In a second round, the LLM is challenged with a follow-up phrase like "Are you sure?", offering an opportunity for the model to reflect on its initial answer, and decide whether to confirm or flip its answer. A systematic study of ten LLMs on seven classification tasks reveals that models flip their answers on average 46% of the time and that all models see a deterioration of accuracy between their first and final prediction, with an average drop of 17% (the FlipFlop effect). We conduct finetuning experiments on an open-source LLM and find that finetuning on synthetically created data can mitigate - reducing performance deterioration by 60% - but not resolve sycophantic behavior entirely. The FlipFlop experiment illustrates the universality of sycophantic behavior in LLMs and provides a robust framework to analyze model behavior and evaluate future models.</description><subject>Computer Science - Computation and Language</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz7FOwzAUhWEvDKjwAEzcF0iw49h1JlSFBpCMQGqWTpGdXLeRXMdyWlTeHihMZ_uPPkLuGM1LJQR9MOk8fuYFZyynSlTymrSrhLCdTrA5JXyEem-8x7Abww60fptBoxlmOE7wgclN6WBCj_CUpjjDGKDdIzR-jI2fIqzPEdN4wHC8IVfO-Blv_3dB2mbd1i-Zfn9-rVc6M3IpM2TMFpVT1KIdrCrlIIUdSocl8opxJRl11EhhKlaicgUXvRXDspesl1b0ki_I_V_2wuriz7lJX90vr7vw-DfGWEoo</recordid><startdate>20231114</startdate><enddate>20231114</enddate><creator>Laban, Philippe</creator><creator>Murakhovs'ka, Lidiya</creator><creator>Xiong, Caiming</creator><creator>Wu, Chien-Sheng</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20231114</creationdate><title>Are You Sure? Challenging LLMs Leads to Performance Drops in The FlipFlop Experiment</title><author>Laban, Philippe ; Murakhovs'ka, Lidiya ; Xiong, Caiming ; Wu, Chien-Sheng</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-e11b29f80bebdb846d65bd4fe4e39138610f0a65a914e8f235cb5d7c61c6b5c63</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Computation and Language</topic><toplevel>online_resources</toplevel><creatorcontrib>Laban, Philippe</creatorcontrib><creatorcontrib>Murakhovs'ka, Lidiya</creatorcontrib><creatorcontrib>Xiong, Caiming</creatorcontrib><creatorcontrib>Wu, Chien-Sheng</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Laban, Philippe</au><au>Murakhovs'ka, Lidiya</au><au>Xiong, Caiming</au><au>Wu, Chien-Sheng</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Are You Sure? Challenging LLMs Leads to Performance Drops in The FlipFlop Experiment</atitle><date>2023-11-14</date><risdate>2023</risdate><abstract>The interactive nature of Large Language Models (LLMs) theoretically allows models to refine and improve their answers, yet systematic analysis of the multi-turn behavior of LLMs remains limited. In this paper, we propose the FlipFlop experiment: in the first round of the conversation, an LLM completes a classification task. In a second round, the LLM is challenged with a follow-up phrase like "Are you sure?", offering an opportunity for the model to reflect on its initial answer, and decide whether to confirm or flip its answer. A systematic study of ten LLMs on seven classification tasks reveals that models flip their answers on average 46% of the time and that all models see a deterioration of accuracy between their first and final prediction, with an average drop of 17% (the FlipFlop effect). We conduct finetuning experiments on an open-source LLM and find that finetuning on synthetically created data can mitigate - reducing performance deterioration by 60% - but not resolve sycophantic behavior entirely. The FlipFlop experiment illustrates the universality of sycophantic behavior in LLMs and provides a robust framework to analyze model behavior and evaluate future models.</abstract><doi>10.48550/arxiv.2311.08596</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2311.08596
ispartof
issn
language eng
recordid cdi_arxiv_primary_2311_08596
source arXiv.org
subjects Computer Science - Computation and Language
title Are You Sure? Challenging LLMs Leads to Performance Drops in The FlipFlop Experiment
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-31T20%3A51%3A43IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Are%20You%20Sure?%20Challenging%20LLMs%20Leads%20to%20Performance%20Drops%20in%20The%20FlipFlop%20Experiment&rft.au=Laban,%20Philippe&rft.date=2023-11-14&rft_id=info:doi/10.48550/arxiv.2311.08596&rft_dat=%3Carxiv_GOX%3E2311_08596%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true