Large Language Models Overcome the Machine Penalty When Acting Fairly but Not When Acting Selfishly or Altruistically

In social dilemmas where the collective and self-interests are at odds, people typically cooperate less with machines than with fellow humans, a phenomenon termed the machine penalty. Overcoming this penalty is critical for successful human-machine collectives, yet current solutions often involve et...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Wang, Zhen, Song, Ruiqi, Shen, Chen, Yin, Shiya, Song, Zhao, Battu, Balaraju, Shi, Lei, Jia, Danyang, Rahwan, Talal, Hu, Shuyue
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Wang, Zhen
Song, Ruiqi
Shen, Chen
Yin, Shiya
Song, Zhao
Battu, Balaraju
Shi, Lei
Jia, Danyang
Rahwan, Talal
Hu, Shuyue
description In social dilemmas where the collective and self-interests are at odds, people typically cooperate less with machines than with fellow humans, a phenomenon termed the machine penalty. Overcoming this penalty is critical for successful human-machine collectives, yet current solutions often involve ethically-questionable tactics, like concealing machines' non-human nature. In this study, with 1,152 participants, we explore the possibility of closing this research question by using Large Language Models (LLMs), in scenarios where communication is possible between interacting parties. We design three types of LLMs: (i) Cooperative, aiming to assist its human associate; (ii) Selfish, focusing solely on maximizing its self-interest; and (iii) Fair, balancing its own and collective interest, while slightly prioritizing self-interest. Our findings reveal that, when interacting with humans, fair LLMs are able to induce cooperation levels comparable to those observed in human-human interactions, even when their non-human nature is fully disclosed. In contrast, selfish and cooperative LLMs fail to achieve this goal. Post-experiment analysis shows that all three types of LLMs succeed in forming mutual cooperation agreements with humans, yet only fair LLMs, which occasionally break their promises, are capable of instilling a perception among humans that cooperating with them is the social norm, and eliciting positive views on their trustworthiness, mindfulness, intelligence, and communication quality. Our findings suggest that for effective human-machine cooperation, bot manufacturers should avoid designing machines with mere rational decision-making or a sole focus on assisting humans. Instead, they should design machines capable of judiciously balancing their own interest and the interest of humans.
doi_str_mv 10.48550/arxiv.2410.03724
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2410_03724</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2410_03724</sourcerecordid><originalsourceid>FETCH-arxiv_primary_2410_037243</originalsourceid><addsrcrecordid>eNqFjrsOgkAURLexMOoHWHl_QESBaEuMxAIfiSaW5IoXdpNlMfsg8vcisbGymsmcKQ5j06XvhZso8heoX6LxVmE3-MF6FQ6ZS1GXBCmq0mFXDvWDpIFTQzqvKwLLuw1zLhTBmRRK28KNk4I4t0KVkKDQsoW7s3Cs7Q-6kCyE4R2tNcTSaieMFTlK2Y7ZoEBpaPLNEZslu-t2P-8Fs6cWFeo2-4hmvWjw__EGPspJkQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Large Language Models Overcome the Machine Penalty When Acting Fairly but Not When Acting Selfishly or Altruistically</title><source>arXiv.org</source><creator>Wang, Zhen ; Song, Ruiqi ; Shen, Chen ; Yin, Shiya ; Song, Zhao ; Battu, Balaraju ; Shi, Lei ; Jia, Danyang ; Rahwan, Talal ; Hu, Shuyue</creator><creatorcontrib>Wang, Zhen ; Song, Ruiqi ; Shen, Chen ; Yin, Shiya ; Song, Zhao ; Battu, Balaraju ; Shi, Lei ; Jia, Danyang ; Rahwan, Talal ; Hu, Shuyue</creatorcontrib><description>In social dilemmas where the collective and self-interests are at odds, people typically cooperate less with machines than with fellow humans, a phenomenon termed the machine penalty. Overcoming this penalty is critical for successful human-machine collectives, yet current solutions often involve ethically-questionable tactics, like concealing machines' non-human nature. In this study, with 1,152 participants, we explore the possibility of closing this research question by using Large Language Models (LLMs), in scenarios where communication is possible between interacting parties. We design three types of LLMs: (i) Cooperative, aiming to assist its human associate; (ii) Selfish, focusing solely on maximizing its self-interest; and (iii) Fair, balancing its own and collective interest, while slightly prioritizing self-interest. Our findings reveal that, when interacting with humans, fair LLMs are able to induce cooperation levels comparable to those observed in human-human interactions, even when their non-human nature is fully disclosed. In contrast, selfish and cooperative LLMs fail to achieve this goal. Post-experiment analysis shows that all three types of LLMs succeed in forming mutual cooperation agreements with humans, yet only fair LLMs, which occasionally break their promises, are capable of instilling a perception among humans that cooperating with them is the social norm, and eliciting positive views on their trustworthiness, mindfulness, intelligence, and communication quality. Our findings suggest that for effective human-machine cooperation, bot manufacturers should avoid designing machines with mere rational decision-making or a sole focus on assisting humans. Instead, they should design machines capable of judiciously balancing their own interest and the interest of humans.</description><identifier>DOI: 10.48550/arxiv.2410.03724</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Computer Science and Game Theory ; Computer Science - Human-Computer Interaction ; Quantitative Finance - Economics</subject><creationdate>2024-09</creationdate><rights>http://creativecommons.org/licenses/by-nc-sa/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2410.03724$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2410.03724$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Wang, Zhen</creatorcontrib><creatorcontrib>Song, Ruiqi</creatorcontrib><creatorcontrib>Shen, Chen</creatorcontrib><creatorcontrib>Yin, Shiya</creatorcontrib><creatorcontrib>Song, Zhao</creatorcontrib><creatorcontrib>Battu, Balaraju</creatorcontrib><creatorcontrib>Shi, Lei</creatorcontrib><creatorcontrib>Jia, Danyang</creatorcontrib><creatorcontrib>Rahwan, Talal</creatorcontrib><creatorcontrib>Hu, Shuyue</creatorcontrib><title>Large Language Models Overcome the Machine Penalty When Acting Fairly but Not When Acting Selfishly or Altruistically</title><description>In social dilemmas where the collective and self-interests are at odds, people typically cooperate less with machines than with fellow humans, a phenomenon termed the machine penalty. Overcoming this penalty is critical for successful human-machine collectives, yet current solutions often involve ethically-questionable tactics, like concealing machines' non-human nature. In this study, with 1,152 participants, we explore the possibility of closing this research question by using Large Language Models (LLMs), in scenarios where communication is possible between interacting parties. We design three types of LLMs: (i) Cooperative, aiming to assist its human associate; (ii) Selfish, focusing solely on maximizing its self-interest; and (iii) Fair, balancing its own and collective interest, while slightly prioritizing self-interest. Our findings reveal that, when interacting with humans, fair LLMs are able to induce cooperation levels comparable to those observed in human-human interactions, even when their non-human nature is fully disclosed. In contrast, selfish and cooperative LLMs fail to achieve this goal. Post-experiment analysis shows that all three types of LLMs succeed in forming mutual cooperation agreements with humans, yet only fair LLMs, which occasionally break their promises, are capable of instilling a perception among humans that cooperating with them is the social norm, and eliciting positive views on their trustworthiness, mindfulness, intelligence, and communication quality. Our findings suggest that for effective human-machine cooperation, bot manufacturers should avoid designing machines with mere rational decision-making or a sole focus on assisting humans. Instead, they should design machines capable of judiciously balancing their own interest and the interest of humans.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Computer Science and Game Theory</subject><subject>Computer Science - Human-Computer Interaction</subject><subject>Quantitative Finance - Economics</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNqFjrsOgkAURLexMOoHWHl_QESBaEuMxAIfiSaW5IoXdpNlMfsg8vcisbGymsmcKQ5j06XvhZso8heoX6LxVmE3-MF6FQ6ZS1GXBCmq0mFXDvWDpIFTQzqvKwLLuw1zLhTBmRRK28KNk4I4t0KVkKDQsoW7s3Cs7Q-6kCyE4R2tNcTSaieMFTlK2Y7ZoEBpaPLNEZslu-t2P-8Fs6cWFeo2-4hmvWjw__EGPspJkQ</recordid><startdate>20240929</startdate><enddate>20240929</enddate><creator>Wang, Zhen</creator><creator>Song, Ruiqi</creator><creator>Shen, Chen</creator><creator>Yin, Shiya</creator><creator>Song, Zhao</creator><creator>Battu, Balaraju</creator><creator>Shi, Lei</creator><creator>Jia, Danyang</creator><creator>Rahwan, Talal</creator><creator>Hu, Shuyue</creator><scope>ADEOX</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20240929</creationdate><title>Large Language Models Overcome the Machine Penalty When Acting Fairly but Not When Acting Selfishly or Altruistically</title><author>Wang, Zhen ; Song, Ruiqi ; Shen, Chen ; Yin, Shiya ; Song, Zhao ; Battu, Balaraju ; Shi, Lei ; Jia, Danyang ; Rahwan, Talal ; Hu, Shuyue</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-arxiv_primary_2410_037243</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Computer Science and Game Theory</topic><topic>Computer Science - Human-Computer Interaction</topic><topic>Quantitative Finance - Economics</topic><toplevel>online_resources</toplevel><creatorcontrib>Wang, Zhen</creatorcontrib><creatorcontrib>Song, Ruiqi</creatorcontrib><creatorcontrib>Shen, Chen</creatorcontrib><creatorcontrib>Yin, Shiya</creatorcontrib><creatorcontrib>Song, Zhao</creatorcontrib><creatorcontrib>Battu, Balaraju</creatorcontrib><creatorcontrib>Shi, Lei</creatorcontrib><creatorcontrib>Jia, Danyang</creatorcontrib><creatorcontrib>Rahwan, Talal</creatorcontrib><creatorcontrib>Hu, Shuyue</creatorcontrib><collection>arXiv Economics</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Wang, Zhen</au><au>Song, Ruiqi</au><au>Shen, Chen</au><au>Yin, Shiya</au><au>Song, Zhao</au><au>Battu, Balaraju</au><au>Shi, Lei</au><au>Jia, Danyang</au><au>Rahwan, Talal</au><au>Hu, Shuyue</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Large Language Models Overcome the Machine Penalty When Acting Fairly but Not When Acting Selfishly or Altruistically</atitle><date>2024-09-29</date><risdate>2024</risdate><abstract>In social dilemmas where the collective and self-interests are at odds, people typically cooperate less with machines than with fellow humans, a phenomenon termed the machine penalty. Overcoming this penalty is critical for successful human-machine collectives, yet current solutions often involve ethically-questionable tactics, like concealing machines' non-human nature. In this study, with 1,152 participants, we explore the possibility of closing this research question by using Large Language Models (LLMs), in scenarios where communication is possible between interacting parties. We design three types of LLMs: (i) Cooperative, aiming to assist its human associate; (ii) Selfish, focusing solely on maximizing its self-interest; and (iii) Fair, balancing its own and collective interest, while slightly prioritizing self-interest. Our findings reveal that, when interacting with humans, fair LLMs are able to induce cooperation levels comparable to those observed in human-human interactions, even when their non-human nature is fully disclosed. In contrast, selfish and cooperative LLMs fail to achieve this goal. Post-experiment analysis shows that all three types of LLMs succeed in forming mutual cooperation agreements with humans, yet only fair LLMs, which occasionally break their promises, are capable of instilling a perception among humans that cooperating with them is the social norm, and eliciting positive views on their trustworthiness, mindfulness, intelligence, and communication quality. Our findings suggest that for effective human-machine cooperation, bot manufacturers should avoid designing machines with mere rational decision-making or a sole focus on assisting humans. Instead, they should design machines capable of judiciously balancing their own interest and the interest of humans.</abstract><doi>10.48550/arxiv.2410.03724</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2410.03724
ispartof
issn
language eng
recordid cdi_arxiv_primary_2410_03724
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Computer Science and Game Theory
Computer Science - Human-Computer Interaction
Quantitative Finance - Economics
title Large Language Models Overcome the Machine Penalty When Acting Fairly but Not When Acting Selfishly or Altruistically
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-07T09%3A38%3A20IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Large%20Language%20Models%20Overcome%20the%20Machine%20Penalty%20When%20Acting%20Fairly%20but%20Not%20When%20Acting%20Selfishly%20or%20Altruistically&rft.au=Wang,%20Zhen&rft.date=2024-09-29&rft_id=info:doi/10.48550/arxiv.2410.03724&rft_dat=%3Carxiv_GOX%3E2410_03724%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true