Confounding-Robust Policy Improvement with Human-AI Teams

Human-AI collaboration has the potential to transform various domains by leveraging the complementary strengths of human experts and Artificial Intelligence (AI) systems. However, unobserved confounding can undermine the effectiveness of this collaboration, leading to biased and unreliable outcomes....

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Gao, Ruijiang, Yin, Mingzhang
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Gao, Ruijiang
Yin, Mingzhang
description Human-AI collaboration has the potential to transform various domains by leveraging the complementary strengths of human experts and Artificial Intelligence (AI) systems. However, unobserved confounding can undermine the effectiveness of this collaboration, leading to biased and unreliable outcomes. In this paper, we propose a novel solution to address unobserved confounding in human-AI collaboration by employing the marginal sensitivity model (MSM). Our approach combines domain expertise with AI-driven statistical modeling to account for potential confounders that may otherwise remain hidden. We present a deferral collaboration framework for incorporating the MSM into policy learning from observational data, enabling the system to control for the influence of unobserved confounding factors. In addition, we propose a personalized deferral collaboration system to leverage the diverse expertise of different human decision-makers. By adjusting for potential biases, our proposed solution enhances the robustness and reliability of collaborative outcomes. The empirical and theoretical analyses demonstrate the efficacy of our approach in mitigating unobserved confounding and improving the overall performance of human-AI collaborations.
doi_str_mv 10.48550/arxiv.2310.08824
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2310_08824</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2310_08824</sourcerecordid><originalsourceid>FETCH-LOGICAL-a674-d77418a5e6b815d5300376abb8f576f64a394d5b25ee8b08ecb4cdccfb5b36463</originalsourceid><addsrcrecordid>eNotj8tKAzEYRrPpQqoP4Mq8QGpmcu2yDLUdKCgy--HPTQNNUubS2rdXq6sD3-LwHYQeK7riWgj6DMNXPK9q9jNQrWt-h9ZNyaHM2cX8Qd6LmccJv5VjtFfcptNQzj75POFLnD7xfk6QyabFnYc03qNFgOPoH_65RN3Ltmv25PC6a5vNgYBUnDileKVBeGl0JZxglDIlwRgdhJJBcmBr7oSphffaUO2t4dZZG4wwTHLJlujpT3v73p-GmGC49r8N_a2BfQPWeEFV</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Confounding-Robust Policy Improvement with Human-AI Teams</title><source>arXiv.org</source><creator>Gao, Ruijiang ; Yin, Mingzhang</creator><creatorcontrib>Gao, Ruijiang ; Yin, Mingzhang</creatorcontrib><description>Human-AI collaboration has the potential to transform various domains by leveraging the complementary strengths of human experts and Artificial Intelligence (AI) systems. However, unobserved confounding can undermine the effectiveness of this collaboration, leading to biased and unreliable outcomes. In this paper, we propose a novel solution to address unobserved confounding in human-AI collaboration by employing the marginal sensitivity model (MSM). Our approach combines domain expertise with AI-driven statistical modeling to account for potential confounders that may otherwise remain hidden. We present a deferral collaboration framework for incorporating the MSM into policy learning from observational data, enabling the system to control for the influence of unobserved confounding factors. In addition, we propose a personalized deferral collaboration system to leverage the diverse expertise of different human decision-makers. By adjusting for potential biases, our proposed solution enhances the robustness and reliability of collaborative outcomes. The empirical and theoretical analyses demonstrate the efficacy of our approach in mitigating unobserved confounding and improving the overall performance of human-AI collaborations.</description><identifier>DOI: 10.48550/arxiv.2310.08824</identifier><language>eng</language><subject>Computer Science - Human-Computer Interaction ; Statistics - Machine Learning</subject><creationdate>2023-10</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,886</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2310.08824$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2310.08824$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Gao, Ruijiang</creatorcontrib><creatorcontrib>Yin, Mingzhang</creatorcontrib><title>Confounding-Robust Policy Improvement with Human-AI Teams</title><description>Human-AI collaboration has the potential to transform various domains by leveraging the complementary strengths of human experts and Artificial Intelligence (AI) systems. However, unobserved confounding can undermine the effectiveness of this collaboration, leading to biased and unreliable outcomes. In this paper, we propose a novel solution to address unobserved confounding in human-AI collaboration by employing the marginal sensitivity model (MSM). Our approach combines domain expertise with AI-driven statistical modeling to account for potential confounders that may otherwise remain hidden. We present a deferral collaboration framework for incorporating the MSM into policy learning from observational data, enabling the system to control for the influence of unobserved confounding factors. In addition, we propose a personalized deferral collaboration system to leverage the diverse expertise of different human decision-makers. By adjusting for potential biases, our proposed solution enhances the robustness and reliability of collaborative outcomes. The empirical and theoretical analyses demonstrate the efficacy of our approach in mitigating unobserved confounding and improving the overall performance of human-AI collaborations.</description><subject>Computer Science - Human-Computer Interaction</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj8tKAzEYRrPpQqoP4Mq8QGpmcu2yDLUdKCgy--HPTQNNUubS2rdXq6sD3-LwHYQeK7riWgj6DMNXPK9q9jNQrWt-h9ZNyaHM2cX8Qd6LmccJv5VjtFfcptNQzj75POFLnD7xfk6QyabFnYc03qNFgOPoH_65RN3Ltmv25PC6a5vNgYBUnDileKVBeGl0JZxglDIlwRgdhJJBcmBr7oSphffaUO2t4dZZG4wwTHLJlujpT3v73p-GmGC49r8N_a2BfQPWeEFV</recordid><startdate>20231012</startdate><enddate>20231012</enddate><creator>Gao, Ruijiang</creator><creator>Yin, Mingzhang</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20231012</creationdate><title>Confounding-Robust Policy Improvement with Human-AI Teams</title><author>Gao, Ruijiang ; Yin, Mingzhang</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a674-d77418a5e6b815d5300376abb8f576f64a394d5b25ee8b08ecb4cdccfb5b36463</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Computer Science - Human-Computer Interaction</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Gao, Ruijiang</creatorcontrib><creatorcontrib>Yin, Mingzhang</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Gao, Ruijiang</au><au>Yin, Mingzhang</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Confounding-Robust Policy Improvement with Human-AI Teams</atitle><date>2023-10-12</date><risdate>2023</risdate><abstract>Human-AI collaboration has the potential to transform various domains by leveraging the complementary strengths of human experts and Artificial Intelligence (AI) systems. However, unobserved confounding can undermine the effectiveness of this collaboration, leading to biased and unreliable outcomes. In this paper, we propose a novel solution to address unobserved confounding in human-AI collaboration by employing the marginal sensitivity model (MSM). Our approach combines domain expertise with AI-driven statistical modeling to account for potential confounders that may otherwise remain hidden. We present a deferral collaboration framework for incorporating the MSM into policy learning from observational data, enabling the system to control for the influence of unobserved confounding factors. In addition, we propose a personalized deferral collaboration system to leverage the diverse expertise of different human decision-makers. By adjusting for potential biases, our proposed solution enhances the robustness and reliability of collaborative outcomes. The empirical and theoretical analyses demonstrate the efficacy of our approach in mitigating unobserved confounding and improving the overall performance of human-AI collaborations.</abstract><doi>10.48550/arxiv.2310.08824</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2310.08824
ispartof
issn
language eng
recordid cdi_arxiv_primary_2310_08824
source arXiv.org
subjects Computer Science - Human-Computer Interaction
Statistics - Machine Learning
title Confounding-Robust Policy Improvement with Human-AI Teams
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-12T16%3A34%3A48IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Confounding-Robust%20Policy%20Improvement%20with%20Human-AI%20Teams&rft.au=Gao,%20Ruijiang&rft.date=2023-10-12&rft_id=info:doi/10.48550/arxiv.2310.08824&rft_dat=%3Carxiv_GOX%3E2310_08824%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true