Fairness and Accuracy Under Domain Generalization

As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in trai...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:ArXiv.org 2023-01
Hauptverfasser: Pham, Thai-Hoang, Zhang, Xueru, Zhang, Ping
Format: Artikel
Sprache:eng
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title ArXiv.org
container_volume
creator Pham, Thai-Hoang
Zhang, Xueru
Zhang, Ping
description As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in training and deployment are identical. Unfortunately, this is commonly violated in practice and a model that is fair during training may lead to an unexpected outcome during its deployment. Although the problem of designing robust ML models under dataset shifts has been widely studied, most existing works focus only on the transfer of accuracy. In this paper, we study the transfer of both fairness and accuracy under domain generalization where the data at test time may be sampled from - - domains. We first develop theoretical bounds on the unfairness and expected loss at deployment, and then derive sufficient conditions under which fairness and accuracy can be perfectly transferred via invariant representation learning. Guided by this, we design a learning algorithm such that fair ML models learned with training data still have high fairness and accuracy when deployment environments change. Experiments on real-world data validate the proposed algorithm. Model implementation is available at https://github.com/pth1993/FATDM.
format Article
fullrecord <record><control><sourceid>proquest_pubme</sourceid><recordid>TN_cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_10246117</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2824684141</sourcerecordid><originalsourceid>FETCH-LOGICAL-p1127-94213c44de350f5df0f525cbbe056d3e92724ec89d1b4972042277f1b7f82ef83</originalsourceid><addsrcrecordid>eNpVkEFLAzEQhYMottT-Bdmjl4XMJNvsnqRUq0LBiz2HbDKrkd1sTbpC_fUuWKVeZgbe43uPOWNTFALyUiKen9wTNk_pnXOOC4VFIS7ZRCisUCqYMlgbHwOllJngsqW1QzT2kG2Do5jd9Z3xIXugQNG0_svsfR-u2EVj2kTz456x7fr-ZfWYb54fnlbLTb4DQJVXEkFYKR2JgjeFa8aBha1r4sXCCapQoSRbVg5qWSnkY1OlGqhVUyI1pZix2x_ubqg7cpbCfiyhd9F3Jh50b7z-rwT_pl_7Tw0c5QJAjYSbIyH2HwOlve58stS2JlA_JI3laCwlSBit16dhfym_jxLfxRBlaQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2824684141</pqid></control><display><type>article</type><title>Fairness and Accuracy Under Domain Generalization</title><source>Free E- Journals</source><creator>Pham, Thai-Hoang ; Zhang, Xueru ; Zhang, Ping</creator><creatorcontrib>Pham, Thai-Hoang ; Zhang, Xueru ; Zhang, Ping</creatorcontrib><description>As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in training and deployment are identical. Unfortunately, this is commonly violated in practice and a model that is fair during training may lead to an unexpected outcome during its deployment. Although the problem of designing robust ML models under dataset shifts has been widely studied, most existing works focus only on the transfer of accuracy. In this paper, we study the transfer of both fairness and accuracy under domain generalization where the data at test time may be sampled from - - domains. We first develop theoretical bounds on the unfairness and expected loss at deployment, and then derive sufficient conditions under which fairness and accuracy can be perfectly transferred via invariant representation learning. Guided by this, we design a learning algorithm such that fair ML models learned with training data still have high fairness and accuracy when deployment environments change. Experiments on real-world data validate the proposed algorithm. Model implementation is available at https://github.com/pth1993/FATDM.</description><identifier>ISSN: 2331-8422</identifier><identifier>EISSN: 2331-8422</identifier><identifier>PMID: 37292471</identifier><language>eng</language><publisher>United States: Cornell University</publisher><ispartof>ArXiv.org, 2023-01</ispartof><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>230,314,780,784,885</link.rule.ids><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/37292471$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Pham, Thai-Hoang</creatorcontrib><creatorcontrib>Zhang, Xueru</creatorcontrib><creatorcontrib>Zhang, Ping</creatorcontrib><title>Fairness and Accuracy Under Domain Generalization</title><title>ArXiv.org</title><addtitle>ArXiv</addtitle><description>As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in training and deployment are identical. Unfortunately, this is commonly violated in practice and a model that is fair during training may lead to an unexpected outcome during its deployment. Although the problem of designing robust ML models under dataset shifts has been widely studied, most existing works focus only on the transfer of accuracy. In this paper, we study the transfer of both fairness and accuracy under domain generalization where the data at test time may be sampled from - - domains. We first develop theoretical bounds on the unfairness and expected loss at deployment, and then derive sufficient conditions under which fairness and accuracy can be perfectly transferred via invariant representation learning. Guided by this, we design a learning algorithm such that fair ML models learned with training data still have high fairness and accuracy when deployment environments change. Experiments on real-world data validate the proposed algorithm. Model implementation is available at https://github.com/pth1993/FATDM.</description><issn>2331-8422</issn><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpVkEFLAzEQhYMottT-Bdmjl4XMJNvsnqRUq0LBiz2HbDKrkd1sTbpC_fUuWKVeZgbe43uPOWNTFALyUiKen9wTNk_pnXOOC4VFIS7ZRCisUCqYMlgbHwOllJngsqW1QzT2kG2Do5jd9Z3xIXugQNG0_svsfR-u2EVj2kTz456x7fr-ZfWYb54fnlbLTb4DQJVXEkFYKR2JgjeFa8aBha1r4sXCCapQoSRbVg5qWSnkY1OlGqhVUyI1pZix2x_ubqg7cpbCfiyhd9F3Jh50b7z-rwT_pl_7Tw0c5QJAjYSbIyH2HwOlve58stS2JlA_JI3laCwlSBit16dhfym_jxLfxRBlaQ</recordid><startdate>20230130</startdate><enddate>20230130</enddate><creator>Pham, Thai-Hoang</creator><creator>Zhang, Xueru</creator><creator>Zhang, Ping</creator><general>Cornell University</general><scope>NPM</scope><scope>7X8</scope><scope>5PM</scope></search><sort><creationdate>20230130</creationdate><title>Fairness and Accuracy Under Domain Generalization</title><author>Pham, Thai-Hoang ; Zhang, Xueru ; Zhang, Ping</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-p1127-94213c44de350f5df0f525cbbe056d3e92724ec89d1b4972042277f1b7f82ef83</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><toplevel>online_resources</toplevel><creatorcontrib>Pham, Thai-Hoang</creatorcontrib><creatorcontrib>Zhang, Xueru</creatorcontrib><creatorcontrib>Zhang, Ping</creatorcontrib><collection>PubMed</collection><collection>MEDLINE - Academic</collection><collection>PubMed Central (Full Participant titles)</collection><jtitle>ArXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Pham, Thai-Hoang</au><au>Zhang, Xueru</au><au>Zhang, Ping</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Fairness and Accuracy Under Domain Generalization</atitle><jtitle>ArXiv.org</jtitle><addtitle>ArXiv</addtitle><date>2023-01-30</date><risdate>2023</risdate><issn>2331-8422</issn><eissn>2331-8422</eissn><abstract>As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in training and deployment are identical. Unfortunately, this is commonly violated in practice and a model that is fair during training may lead to an unexpected outcome during its deployment. Although the problem of designing robust ML models under dataset shifts has been widely studied, most existing works focus only on the transfer of accuracy. In this paper, we study the transfer of both fairness and accuracy under domain generalization where the data at test time may be sampled from - - domains. We first develop theoretical bounds on the unfairness and expected loss at deployment, and then derive sufficient conditions under which fairness and accuracy can be perfectly transferred via invariant representation learning. Guided by this, we design a learning algorithm such that fair ML models learned with training data still have high fairness and accuracy when deployment environments change. Experiments on real-world data validate the proposed algorithm. Model implementation is available at https://github.com/pth1993/FATDM.</abstract><cop>United States</cop><pub>Cornell University</pub><pmid>37292471</pmid><oa>free_for_read</oa></addata></record>
fulltext fulltext
identifier ISSN: 2331-8422
ispartof ArXiv.org, 2023-01
issn 2331-8422
2331-8422
language eng
recordid cdi_pubmedcentral_primary_oai_pubmedcentral_nih_gov_10246117
source Free E- Journals
title Fairness and Accuracy Under Domain Generalization
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T13%3A37%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_pubme&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Fairness%20and%20Accuracy%20Under%20Domain%20Generalization&rft.jtitle=ArXiv.org&rft.au=Pham,%20Thai-Hoang&rft.date=2023-01-30&rft.issn=2331-8422&rft.eissn=2331-8422&rft_id=info:doi/&rft_dat=%3Cproquest_pubme%3E2824684141%3C/proquest_pubme%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2824684141&rft_id=info:pmid/37292471&rfr_iscdi=true