Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals

There has been a recent resurgence of interest in explainable artificial intelligence (XAI) that aims to reduce the opaqueness of AI-based decision-making systems, allowing humans to scrutinize and trust them. Prior work in this context has focused on the attribution of responsibility for an algorit...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Galhotra, Sainyam, Pradhan, Romila, Salimi, Babak
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Galhotra, Sainyam
Pradhan, Romila
Salimi, Babak
description There has been a recent resurgence of interest in explainable artificial intelligence (XAI) that aims to reduce the opaqueness of AI-based decision-making systems, allowing humans to scrutinize and trust them. Prior work in this context has focused on the attribution of responsibility for an algorithm's decisions to its inputs wherein responsibility is typically approached as a purely associational concept. In this paper, we propose a principled causality-based approach for explaining black-box decision-making systems that addresses limitations of existing methods in XAI. At the core of our framework lies probabilistic contrastive counterfactuals, a concept that can be traced back to philosophical, cognitive, and social foundations of theories on how humans generate and select explanations. We show how such counterfactuals can quantify the direct and indirect influences of a variable on decisions made by an algorithm, and provide actionable recourse for individuals negatively affected by the algorithm's decision. Unlike prior work, our system, LEWIS: (1)can compute provably effective explanations and recourse at local, global and contextual levels (2)is designed to work with users with varying levels of background knowledge of the underlying causal model and (3)makes no assumptions about the internals of an algorithmic system except for the availability of its input-output data. We empirically evaluate LEWIS on three real-world datasets and show that it generates human-understandable explanations that improve upon state-of-the-art approaches in XAI, including the popular LIME and SHAP. Experiments on synthetic data further demonstrate the correctness of LEWIS's explanations and the scalability of its recourse algorithm.
doi_str_mv 10.48550/arxiv.2103.11972
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2103_11972</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2103_11972</sourcerecordid><originalsourceid>FETCH-LOGICAL-a672-eb71303c2592f283cdc16e1b2b62aa0cfd773bb05591abfb1169bc7f6f966f553</originalsourceid><addsrcrecordid>eNotj71OwzAURr0woMIDMJEXSPC1sV2PbVR-pEplKHN0rxsXCzepHLcKb09amM6RPumTDmMPwKvnuVL8CdMYzpUALisAa8Qt26zGY8TQhW5fLCO673LZj8Ui7vsU8tdhKD6Hy_SRekIKMQw5uKLuu5xw0nM7-anLbfLo8gnjcMdu_IT2_p8ztn1Zbeu3cr15fa8X6xK1EWVLBiSXTigrvJhLt3OgWyBBWiBy53fGSCKulAUkTwDakjNee6u1V0rO2OPf7bWoOaZwwPTTXMqaa5n8BXYvSk0</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals</title><source>arXiv.org</source><creator>Galhotra, Sainyam ; Pradhan, Romila ; Salimi, Babak</creator><creatorcontrib>Galhotra, Sainyam ; Pradhan, Romila ; Salimi, Babak</creatorcontrib><description>There has been a recent resurgence of interest in explainable artificial intelligence (XAI) that aims to reduce the opaqueness of AI-based decision-making systems, allowing humans to scrutinize and trust them. Prior work in this context has focused on the attribution of responsibility for an algorithm's decisions to its inputs wherein responsibility is typically approached as a purely associational concept. In this paper, we propose a principled causality-based approach for explaining black-box decision-making systems that addresses limitations of existing methods in XAI. At the core of our framework lies probabilistic contrastive counterfactuals, a concept that can be traced back to philosophical, cognitive, and social foundations of theories on how humans generate and select explanations. We show how such counterfactuals can quantify the direct and indirect influences of a variable on decisions made by an algorithm, and provide actionable recourse for individuals negatively affected by the algorithm's decision. Unlike prior work, our system, LEWIS: (1)can compute provably effective explanations and recourse at local, global and contextual levels (2)is designed to work with users with varying levels of background knowledge of the underlying causal model and (3)makes no assumptions about the internals of an algorithmic system except for the availability of its input-output data. We empirically evaluate LEWIS on three real-world datasets and show that it generates human-understandable explanations that improve upon state-of-the-art approaches in XAI, including the popular LIME and SHAP. Experiments on synthetic data further demonstrate the correctness of LEWIS's explanations and the scalability of its recourse algorithm.</description><identifier>DOI: 10.48550/arxiv.2103.11972</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Databases ; Computer Science - Learning</subject><creationdate>2021-03</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2103.11972$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2103.11972$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Galhotra, Sainyam</creatorcontrib><creatorcontrib>Pradhan, Romila</creatorcontrib><creatorcontrib>Salimi, Babak</creatorcontrib><title>Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals</title><description>There has been a recent resurgence of interest in explainable artificial intelligence (XAI) that aims to reduce the opaqueness of AI-based decision-making systems, allowing humans to scrutinize and trust them. Prior work in this context has focused on the attribution of responsibility for an algorithm's decisions to its inputs wherein responsibility is typically approached as a purely associational concept. In this paper, we propose a principled causality-based approach for explaining black-box decision-making systems that addresses limitations of existing methods in XAI. At the core of our framework lies probabilistic contrastive counterfactuals, a concept that can be traced back to philosophical, cognitive, and social foundations of theories on how humans generate and select explanations. We show how such counterfactuals can quantify the direct and indirect influences of a variable on decisions made by an algorithm, and provide actionable recourse for individuals negatively affected by the algorithm's decision. Unlike prior work, our system, LEWIS: (1)can compute provably effective explanations and recourse at local, global and contextual levels (2)is designed to work with users with varying levels of background knowledge of the underlying causal model and (3)makes no assumptions about the internals of an algorithmic system except for the availability of its input-output data. We empirically evaluate LEWIS on three real-world datasets and show that it generates human-understandable explanations that improve upon state-of-the-art approaches in XAI, including the popular LIME and SHAP. Experiments on synthetic data further demonstrate the correctness of LEWIS's explanations and the scalability of its recourse algorithm.</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Databases</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71OwzAURr0woMIDMJEXSPC1sV2PbVR-pEplKHN0rxsXCzepHLcKb09amM6RPumTDmMPwKvnuVL8CdMYzpUALisAa8Qt26zGY8TQhW5fLCO673LZj8Ui7vsU8tdhKD6Hy_SRekIKMQw5uKLuu5xw0nM7-anLbfLo8gnjcMdu_IT2_p8ztn1Zbeu3cr15fa8X6xK1EWVLBiSXTigrvJhLt3OgWyBBWiBy53fGSCKulAUkTwDakjNee6u1V0rO2OPf7bWoOaZwwPTTXMqaa5n8BXYvSk0</recordid><startdate>20210322</startdate><enddate>20210322</enddate><creator>Galhotra, Sainyam</creator><creator>Pradhan, Romila</creator><creator>Salimi, Babak</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20210322</creationdate><title>Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals</title><author>Galhotra, Sainyam ; Pradhan, Romila ; Salimi, Babak</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a672-eb71303c2592f283cdc16e1b2b62aa0cfd773bb05591abfb1169bc7f6f966f553</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Databases</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Galhotra, Sainyam</creatorcontrib><creatorcontrib>Pradhan, Romila</creatorcontrib><creatorcontrib>Salimi, Babak</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Galhotra, Sainyam</au><au>Pradhan, Romila</au><au>Salimi, Babak</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals</atitle><date>2021-03-22</date><risdate>2021</risdate><abstract>There has been a recent resurgence of interest in explainable artificial intelligence (XAI) that aims to reduce the opaqueness of AI-based decision-making systems, allowing humans to scrutinize and trust them. Prior work in this context has focused on the attribution of responsibility for an algorithm's decisions to its inputs wherein responsibility is typically approached as a purely associational concept. In this paper, we propose a principled causality-based approach for explaining black-box decision-making systems that addresses limitations of existing methods in XAI. At the core of our framework lies probabilistic contrastive counterfactuals, a concept that can be traced back to philosophical, cognitive, and social foundations of theories on how humans generate and select explanations. We show how such counterfactuals can quantify the direct and indirect influences of a variable on decisions made by an algorithm, and provide actionable recourse for individuals negatively affected by the algorithm's decision. Unlike prior work, our system, LEWIS: (1)can compute provably effective explanations and recourse at local, global and contextual levels (2)is designed to work with users with varying levels of background knowledge of the underlying causal model and (3)makes no assumptions about the internals of an algorithmic system except for the availability of its input-output data. We empirically evaluate LEWIS on three real-world datasets and show that it generates human-understandable explanations that improve upon state-of-the-art approaches in XAI, including the popular LIME and SHAP. Experiments on synthetic data further demonstrate the correctness of LEWIS's explanations and the scalability of its recourse algorithm.</abstract><doi>10.48550/arxiv.2103.11972</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2103.11972
ispartof
issn
language eng
recordid cdi_arxiv_primary_2103_11972
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Databases
Computer Science - Learning
title Explaining Black-Box Algorithms Using Probabilistic Contrastive Counterfactuals
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-19T23%3A49%3A33IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Explaining%20Black-Box%20Algorithms%20Using%20Probabilistic%20Contrastive%20Counterfactuals&rft.au=Galhotra,%20Sainyam&rft.date=2021-03-22&rft_id=info:doi/10.48550/arxiv.2103.11972&rft_dat=%3Carxiv_GOX%3E2103_11972%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true