A Statistical Test for Probabilistic Fairness
Algorithms are now routinely used to make consequential decisions that affect human lives. Examples include college admissions, medical interventions or law enforcement. While algorithms empower us to harness all information hidden in vast amounts of data, they may inadvertently amplify existing bia...
Gespeichert in:
Hauptverfasser: | , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Taskesen, Bahar Blanchet, Jose Kuhn, Daniel Nguyen, Viet Anh |
description | Algorithms are now routinely used to make consequential decisions that affect
human lives. Examples include college admissions, medical interventions or law
enforcement. While algorithms empower us to harness all information hidden in
vast amounts of data, they may inadvertently amplify existing biases in the
available datasets. This concern has sparked increasing interest in fair
machine learning, which aims to quantify and mitigate algorithmic
discrimination. Indeed, machine learning models should undergo intensive tests
to detect algorithmic biases before being deployed at scale. In this paper, we
use ideas from the theory of optimal transport to propose a statistical
hypothesis test for detecting unfair classifiers. Leveraging the geometry of
the feature space, the test statistic quantifies the distance of the empirical
distribution supported on the test samples to the manifold of distributions
that render a pre-trained classifier fair. We develop a rigorous hypothesis
testing mechanism for assessing the probabilistic fairness of any pre-trained
logistic classifier, and we show both theoretically as well as empirically that
the proposed test is asymptotically correct. In addition, the proposed
framework offers interpretability by identifying the most favorable
perturbation of the data so that the given classifier becomes fair. |
doi_str_mv | 10.48550/arxiv.2012.04800 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2012_04800</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2012_04800</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-1281a9098912ec141935bcd452dda7b83fc3e2ca672fae9c7b8b0ed9241a77463</originalsourceid><addsrcrecordid>eNotjssKwjAURLNxIeoHuDI_0HrzqE2WRXyBoGD35SZNIVCtJEX079XqamDmMBxC5gxSqbIMlhie_pFyYDwFqQDGJCnopcfex95bbGnpYk-bLtBz6Awa3w4D3aIPNxfjlIwabKOb_XNCyu2mXO-T42l3WBfHBFc5JIwrhhq00ow7yyTTIjO2lhmva8yNEo0VjtsPyxt02n4qA67WXDLMc7kSE7L43Q6-1T34K4ZX9fWuBm_xBgxyPGU</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Statistical Test for Probabilistic Fairness</title><source>arXiv.org</source><creator>Taskesen, Bahar ; Blanchet, Jose ; Kuhn, Daniel ; Nguyen, Viet Anh</creator><creatorcontrib>Taskesen, Bahar ; Blanchet, Jose ; Kuhn, Daniel ; Nguyen, Viet Anh</creatorcontrib><description>Algorithms are now routinely used to make consequential decisions that affect
human lives. Examples include college admissions, medical interventions or law
enforcement. While algorithms empower us to harness all information hidden in
vast amounts of data, they may inadvertently amplify existing biases in the
available datasets. This concern has sparked increasing interest in fair
machine learning, which aims to quantify and mitigate algorithmic
discrimination. Indeed, machine learning models should undergo intensive tests
to detect algorithmic biases before being deployed at scale. In this paper, we
use ideas from the theory of optimal transport to propose a statistical
hypothesis test for detecting unfair classifiers. Leveraging the geometry of
the feature space, the test statistic quantifies the distance of the empirical
distribution supported on the test samples to the manifold of distributions
that render a pre-trained classifier fair. We develop a rigorous hypothesis
testing mechanism for assessing the probabilistic fairness of any pre-trained
logistic classifier, and we show both theoretically as well as empirically that
the proposed test is asymptotically correct. In addition, the proposed
framework offers interpretability by identifying the most favorable
perturbation of the data so that the given classifier becomes fair.</description><identifier>DOI: 10.48550/arxiv.2012.04800</identifier><language>eng</language><subject>Computer Science - Computers and Society ; Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2020-12</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2012.04800$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2012.04800$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Taskesen, Bahar</creatorcontrib><creatorcontrib>Blanchet, Jose</creatorcontrib><creatorcontrib>Kuhn, Daniel</creatorcontrib><creatorcontrib>Nguyen, Viet Anh</creatorcontrib><title>A Statistical Test for Probabilistic Fairness</title><description>Algorithms are now routinely used to make consequential decisions that affect
human lives. Examples include college admissions, medical interventions or law
enforcement. While algorithms empower us to harness all information hidden in
vast amounts of data, they may inadvertently amplify existing biases in the
available datasets. This concern has sparked increasing interest in fair
machine learning, which aims to quantify and mitigate algorithmic
discrimination. Indeed, machine learning models should undergo intensive tests
to detect algorithmic biases before being deployed at scale. In this paper, we
use ideas from the theory of optimal transport to propose a statistical
hypothesis test for detecting unfair classifiers. Leveraging the geometry of
the feature space, the test statistic quantifies the distance of the empirical
distribution supported on the test samples to the manifold of distributions
that render a pre-trained classifier fair. We develop a rigorous hypothesis
testing mechanism for assessing the probabilistic fairness of any pre-trained
logistic classifier, and we show both theoretically as well as empirically that
the proposed test is asymptotically correct. In addition, the proposed
framework offers interpretability by identifying the most favorable
perturbation of the data so that the given classifier becomes fair.</description><subject>Computer Science - Computers and Society</subject><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2020</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotjssKwjAURLNxIeoHuDI_0HrzqE2WRXyBoGD35SZNIVCtJEX079XqamDmMBxC5gxSqbIMlhie_pFyYDwFqQDGJCnopcfex95bbGnpYk-bLtBz6Awa3w4D3aIPNxfjlIwabKOb_XNCyu2mXO-T42l3WBfHBFc5JIwrhhq00ow7yyTTIjO2lhmva8yNEo0VjtsPyxt02n4qA67WXDLMc7kSE7L43Q6-1T34K4ZX9fWuBm_xBgxyPGU</recordid><startdate>20201208</startdate><enddate>20201208</enddate><creator>Taskesen, Bahar</creator><creator>Blanchet, Jose</creator><creator>Kuhn, Daniel</creator><creator>Nguyen, Viet Anh</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20201208</creationdate><title>A Statistical Test for Probabilistic Fairness</title><author>Taskesen, Bahar ; Blanchet, Jose ; Kuhn, Daniel ; Nguyen, Viet Anh</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-1281a9098912ec141935bcd452dda7b83fc3e2ca672fae9c7b8b0ed9241a77463</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2020</creationdate><topic>Computer Science - Computers and Society</topic><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Taskesen, Bahar</creatorcontrib><creatorcontrib>Blanchet, Jose</creatorcontrib><creatorcontrib>Kuhn, Daniel</creatorcontrib><creatorcontrib>Nguyen, Viet Anh</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Taskesen, Bahar</au><au>Blanchet, Jose</au><au>Kuhn, Daniel</au><au>Nguyen, Viet Anh</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Statistical Test for Probabilistic Fairness</atitle><date>2020-12-08</date><risdate>2020</risdate><abstract>Algorithms are now routinely used to make consequential decisions that affect
human lives. Examples include college admissions, medical interventions or law
enforcement. While algorithms empower us to harness all information hidden in
vast amounts of data, they may inadvertently amplify existing biases in the
available datasets. This concern has sparked increasing interest in fair
machine learning, which aims to quantify and mitigate algorithmic
discrimination. Indeed, machine learning models should undergo intensive tests
to detect algorithmic biases before being deployed at scale. In this paper, we
use ideas from the theory of optimal transport to propose a statistical
hypothesis test for detecting unfair classifiers. Leveraging the geometry of
the feature space, the test statistic quantifies the distance of the empirical
distribution supported on the test samples to the manifold of distributions
that render a pre-trained classifier fair. We develop a rigorous hypothesis
testing mechanism for assessing the probabilistic fairness of any pre-trained
logistic classifier, and we show both theoretically as well as empirically that
the proposed test is asymptotically correct. In addition, the proposed
framework offers interpretability by identifying the most favorable
perturbation of the data so that the given classifier becomes fair.</abstract><doi>10.48550/arxiv.2012.04800</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2012.04800 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2012_04800 |
source | arXiv.org |
subjects | Computer Science - Computers and Society Computer Science - Learning Statistics - Machine Learning |
title | A Statistical Test for Probabilistic Fairness |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-12T14%3A03%3A06IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Statistical%20Test%20for%20Probabilistic%20Fairness&rft.au=Taskesen,%20Bahar&rft.date=2020-12-08&rft_id=info:doi/10.48550/arxiv.2012.04800&rft_dat=%3Carxiv_GOX%3E2012_04800%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |