Fair When Trained, Unfair When Deployed: Observable Fairness Measures are Unstable in Performative Prediction Settings

Many popular algorithmic fairness measures depend on the joint distribution of predictions, outcomes, and a sensitive feature like race or gender. These measures are sensitive to distribution shift: a predictor which is trained to satisfy one of these fairness definitions may become unfair if the di...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Mishler, Alan, Dalmasso, Niccolò
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Mishler, Alan
Dalmasso, Niccolò
description Many popular algorithmic fairness measures depend on the joint distribution of predictions, outcomes, and a sensitive feature like race or gender. These measures are sensitive to distribution shift: a predictor which is trained to satisfy one of these fairness definitions may become unfair if the distribution changes. In performative prediction settings, however, predictors are precisely intended to induce distribution shift. For example, in many applications in criminal justice, healthcare, and consumer finance, the purpose of building a predictor is to reduce the rate of adverse outcomes such as recidivism, hospitalization, or default on a loan. We formalize the effect of such predictors as a type of concept shift-a particular variety of distribution shift-and show both theoretically and via simulated examples how this causes predictors which are fair when they are trained to become unfair when they are deployed. We further show how many of these issues can be avoided by using fairness definitions that depend on counterfactual rather than observable outcomes.
doi_str_mv 10.48550/arxiv.2202.05049
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2202_05049</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2202_05049</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-2811f25ad39b8d87cb6d1861911a54cd9057210a68a337faa148a5a934bfe7153</originalsourceid><addsrcrecordid>eNo9kM9KxDAYxHPxIKsP4Mk8gK3507SJN1ldFVZ2wYrH8rX5ooFuuiS1uG-vW8XTwAy_gRlCLjjLC60Uu4b45adcCCZyplhhTsm0Ah_p2wcGWkfwAe0VfQ3u37zDfT8c0N7QTZswTtD2SI9MwJToM0L6jJgoRPzB0jjHPtAtRjfEHYx-QrqNaH03-iHQFxxHH97TGTlx0Cc8_9MFqVf39fIxW28enpa36wzKymRCc-6EAitNq62uura0XJfccA6q6KxhqhKcQalBysoB8EKDAiOL1mHFlVyQy9_aeXizj34H8dAcD2jmA-Q3oUlWsw</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Fair When Trained, Unfair When Deployed: Observable Fairness Measures are Unstable in Performative Prediction Settings</title><source>arXiv.org</source><creator>Mishler, Alan ; Dalmasso, Niccolò</creator><creatorcontrib>Mishler, Alan ; Dalmasso, Niccolò</creatorcontrib><description>Many popular algorithmic fairness measures depend on the joint distribution of predictions, outcomes, and a sensitive feature like race or gender. These measures are sensitive to distribution shift: a predictor which is trained to satisfy one of these fairness definitions may become unfair if the distribution changes. In performative prediction settings, however, predictors are precisely intended to induce distribution shift. For example, in many applications in criminal justice, healthcare, and consumer finance, the purpose of building a predictor is to reduce the rate of adverse outcomes such as recidivism, hospitalization, or default on a loan. We formalize the effect of such predictors as a type of concept shift-a particular variety of distribution shift-and show both theoretically and via simulated examples how this causes predictors which are fair when they are trained to become unfair when they are deployed. We further show how many of these issues can be avoided by using fairness definitions that depend on counterfactual rather than observable outcomes.</description><identifier>DOI: 10.48550/arxiv.2202.05049</identifier><language>eng</language><subject>Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2022-02</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2202.05049$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2202.05049$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Mishler, Alan</creatorcontrib><creatorcontrib>Dalmasso, Niccolò</creatorcontrib><title>Fair When Trained, Unfair When Deployed: Observable Fairness Measures are Unstable in Performative Prediction Settings</title><description>Many popular algorithmic fairness measures depend on the joint distribution of predictions, outcomes, and a sensitive feature like race or gender. These measures are sensitive to distribution shift: a predictor which is trained to satisfy one of these fairness definitions may become unfair if the distribution changes. In performative prediction settings, however, predictors are precisely intended to induce distribution shift. For example, in many applications in criminal justice, healthcare, and consumer finance, the purpose of building a predictor is to reduce the rate of adverse outcomes such as recidivism, hospitalization, or default on a loan. We formalize the effect of such predictors as a type of concept shift-a particular variety of distribution shift-and show both theoretically and via simulated examples how this causes predictors which are fair when they are trained to become unfair when they are deployed. We further show how many of these issues can be avoided by using fairness definitions that depend on counterfactual rather than observable outcomes.</description><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNo9kM9KxDAYxHPxIKsP4Mk8gK3507SJN1ldFVZ2wYrH8rX5ooFuuiS1uG-vW8XTwAy_gRlCLjjLC60Uu4b45adcCCZyplhhTsm0Ah_p2wcGWkfwAe0VfQ3u37zDfT8c0N7QTZswTtD2SI9MwJToM0L6jJgoRPzB0jjHPtAtRjfEHYx-QrqNaH03-iHQFxxHH97TGTlx0Cc8_9MFqVf39fIxW28enpa36wzKymRCc-6EAitNq62uura0XJfccA6q6KxhqhKcQalBysoB8EKDAiOL1mHFlVyQy9_aeXizj34H8dAcD2jmA-Q3oUlWsw</recordid><startdate>20220210</startdate><enddate>20220210</enddate><creator>Mishler, Alan</creator><creator>Dalmasso, Niccolò</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20220210</creationdate><title>Fair When Trained, Unfair When Deployed: Observable Fairness Measures are Unstable in Performative Prediction Settings</title><author>Mishler, Alan ; Dalmasso, Niccolò</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-2811f25ad39b8d87cb6d1861911a54cd9057210a68a337faa148a5a934bfe7153</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Mishler, Alan</creatorcontrib><creatorcontrib>Dalmasso, Niccolò</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Mishler, Alan</au><au>Dalmasso, Niccolò</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Fair When Trained, Unfair When Deployed: Observable Fairness Measures are Unstable in Performative Prediction Settings</atitle><date>2022-02-10</date><risdate>2022</risdate><abstract>Many popular algorithmic fairness measures depend on the joint distribution of predictions, outcomes, and a sensitive feature like race or gender. These measures are sensitive to distribution shift: a predictor which is trained to satisfy one of these fairness definitions may become unfair if the distribution changes. In performative prediction settings, however, predictors are precisely intended to induce distribution shift. For example, in many applications in criminal justice, healthcare, and consumer finance, the purpose of building a predictor is to reduce the rate of adverse outcomes such as recidivism, hospitalization, or default on a loan. We formalize the effect of such predictors as a type of concept shift-a particular variety of distribution shift-and show both theoretically and via simulated examples how this causes predictors which are fair when they are trained to become unfair when they are deployed. We further show how many of these issues can be avoided by using fairness definitions that depend on counterfactual rather than observable outcomes.</abstract><doi>10.48550/arxiv.2202.05049</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2202.05049
ispartof
issn
language eng
recordid cdi_arxiv_primary_2202_05049
source arXiv.org
subjects Computer Science - Learning
Statistics - Machine Learning
title Fair When Trained, Unfair When Deployed: Observable Fairness Measures are Unstable in Performative Prediction Settings
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-08T00%3A26%3A53IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Fair%20When%20Trained,%20Unfair%20When%20Deployed:%20Observable%20Fairness%20Measures%20are%20Unstable%20in%20Performative%20Prediction%20Settings&rft.au=Mishler,%20Alan&rft.date=2022-02-10&rft_id=info:doi/10.48550/arxiv.2202.05049&rft_dat=%3Carxiv_GOX%3E2202_05049%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true