Mitigating Sybils in Federated Learning Poisoning

Machine learning (ML) over distributed multi-party data is required for a variety of domains. Existing approaches, such as federated learning, collect the outputs computed by a group of devices at a central aggregator and run iterative algorithms to train a globally shared model. Unfortunately, such...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Fung, Clement, Yoon, Chris J. M, Beschastnikh, Ivan
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Fung, Clement
Yoon, Chris J. M
Beschastnikh, Ivan
description Machine learning (ML) over distributed multi-party data is required for a variety of domains. Existing approaches, such as federated learning, collect the outputs computed by a group of devices at a central aggregator and run iterative algorithms to train a globally shared model. Unfortunately, such approaches are susceptible to a variety of attacks, including model poisoning, which is made substantially worse in the presence of sybils. In this paper we first evaluate the vulnerability of federated learning to sybil-based poisoning attacks. We then describe \emph{FoolsGold}, a novel defense to this problem that identifies poisoning sybils based on the diversity of client updates in the distributed learning process. Unlike prior work, our system does not bound the expected number of attackers, requires no auxiliary information outside of the learning process, and makes fewer assumptions about clients and their data. In our evaluation we show that FoolsGold exceeds the capabilities of existing state of the art approaches to countering sybil-based label-flipping and backdoor poisoning attacks. Our results hold for different distributions of client data, varying poisoning targets, and various sybil strategies. Code can be found at: https://github.com/DistributedML/FoolsGold
doi_str_mv 10.48550/arxiv.1808.04866
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1808_04866</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1808_04866</sourcerecordid><originalsourceid>FETCH-LOGICAL-a676-96441e526a3880d0337f7d402cda19295d3cdfc02c9cf348372412be75d23bb33</originalsourceid><addsrcrecordid>eNotjssKwjAURLNxIeoHuLI_0Jrk5tWliC-oKNh9uW1SCWiVtIj-vVZdzQwHhkPIlNFEGCnpHMPTPxJmqEmoMEoNCdv7zp-x8805Or1Kf2kj30RrZ13AztkocxiaHh5vvr31bUwGNV5aN_nniOTrVb7cxtlhs1sushiVVnGqhGBOcoVgDLUUQNfaCsoriyzlqbRQ2br67LSqQRjQXDBeOi0th7IEGJHZ7_brXNyDv2J4Fb178XWHNz1fPZk</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Mitigating Sybils in Federated Learning Poisoning</title><source>arXiv.org</source><creator>Fung, Clement ; Yoon, Chris J. M ; Beschastnikh, Ivan</creator><creatorcontrib>Fung, Clement ; Yoon, Chris J. M ; Beschastnikh, Ivan</creatorcontrib><description>Machine learning (ML) over distributed multi-party data is required for a variety of domains. Existing approaches, such as federated learning, collect the outputs computed by a group of devices at a central aggregator and run iterative algorithms to train a globally shared model. Unfortunately, such approaches are susceptible to a variety of attacks, including model poisoning, which is made substantially worse in the presence of sybils. In this paper we first evaluate the vulnerability of federated learning to sybil-based poisoning attacks. We then describe \emph{FoolsGold}, a novel defense to this problem that identifies poisoning sybils based on the diversity of client updates in the distributed learning process. Unlike prior work, our system does not bound the expected number of attackers, requires no auxiliary information outside of the learning process, and makes fewer assumptions about clients and their data. In our evaluation we show that FoolsGold exceeds the capabilities of existing state of the art approaches to countering sybil-based label-flipping and backdoor poisoning attacks. Our results hold for different distributions of client data, varying poisoning targets, and various sybil strategies. Code can be found at: https://github.com/DistributedML/FoolsGold</description><identifier>DOI: 10.48550/arxiv.1808.04866</identifier><language>eng</language><subject>Computer Science - Cryptography and Security ; Computer Science - Distributed, Parallel, and Cluster Computing ; Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2018-08</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1808.04866$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1808.04866$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Fung, Clement</creatorcontrib><creatorcontrib>Yoon, Chris J. M</creatorcontrib><creatorcontrib>Beschastnikh, Ivan</creatorcontrib><title>Mitigating Sybils in Federated Learning Poisoning</title><description>Machine learning (ML) over distributed multi-party data is required for a variety of domains. Existing approaches, such as federated learning, collect the outputs computed by a group of devices at a central aggregator and run iterative algorithms to train a globally shared model. Unfortunately, such approaches are susceptible to a variety of attacks, including model poisoning, which is made substantially worse in the presence of sybils. In this paper we first evaluate the vulnerability of federated learning to sybil-based poisoning attacks. We then describe \emph{FoolsGold}, a novel defense to this problem that identifies poisoning sybils based on the diversity of client updates in the distributed learning process. Unlike prior work, our system does not bound the expected number of attackers, requires no auxiliary information outside of the learning process, and makes fewer assumptions about clients and their data. In our evaluation we show that FoolsGold exceeds the capabilities of existing state of the art approaches to countering sybil-based label-flipping and backdoor poisoning attacks. Our results hold for different distributions of client data, varying poisoning targets, and various sybil strategies. Code can be found at: https://github.com/DistributedML/FoolsGold</description><subject>Computer Science - Cryptography and Security</subject><subject>Computer Science - Distributed, Parallel, and Cluster Computing</subject><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2018</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotjssKwjAURLNxIeoHuLI_0Jrk5tWliC-oKNh9uW1SCWiVtIj-vVZdzQwHhkPIlNFEGCnpHMPTPxJmqEmoMEoNCdv7zp-x8805Or1Kf2kj30RrZ13AztkocxiaHh5vvr31bUwGNV5aN_nniOTrVb7cxtlhs1sushiVVnGqhGBOcoVgDLUUQNfaCsoriyzlqbRQ2br67LSqQRjQXDBeOi0th7IEGJHZ7_brXNyDv2J4Fb178XWHNz1fPZk</recordid><startdate>20180814</startdate><enddate>20180814</enddate><creator>Fung, Clement</creator><creator>Yoon, Chris J. M</creator><creator>Beschastnikh, Ivan</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20180814</creationdate><title>Mitigating Sybils in Federated Learning Poisoning</title><author>Fung, Clement ; Yoon, Chris J. M ; Beschastnikh, Ivan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a676-96441e526a3880d0337f7d402cda19295d3cdfc02c9cf348372412be75d23bb33</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2018</creationdate><topic>Computer Science - Cryptography and Security</topic><topic>Computer Science - Distributed, Parallel, and Cluster Computing</topic><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Fung, Clement</creatorcontrib><creatorcontrib>Yoon, Chris J. M</creatorcontrib><creatorcontrib>Beschastnikh, Ivan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Fung, Clement</au><au>Yoon, Chris J. M</au><au>Beschastnikh, Ivan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Mitigating Sybils in Federated Learning Poisoning</atitle><date>2018-08-14</date><risdate>2018</risdate><abstract>Machine learning (ML) over distributed multi-party data is required for a variety of domains. Existing approaches, such as federated learning, collect the outputs computed by a group of devices at a central aggregator and run iterative algorithms to train a globally shared model. Unfortunately, such approaches are susceptible to a variety of attacks, including model poisoning, which is made substantially worse in the presence of sybils. In this paper we first evaluate the vulnerability of federated learning to sybil-based poisoning attacks. We then describe \emph{FoolsGold}, a novel defense to this problem that identifies poisoning sybils based on the diversity of client updates in the distributed learning process. Unlike prior work, our system does not bound the expected number of attackers, requires no auxiliary information outside of the learning process, and makes fewer assumptions about clients and their data. In our evaluation we show that FoolsGold exceeds the capabilities of existing state of the art approaches to countering sybil-based label-flipping and backdoor poisoning attacks. Our results hold for different distributions of client data, varying poisoning targets, and various sybil strategies. Code can be found at: https://github.com/DistributedML/FoolsGold</abstract><doi>10.48550/arxiv.1808.04866</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1808.04866
ispartof
issn
language eng
recordid cdi_arxiv_primary_1808_04866
source arXiv.org
subjects Computer Science - Cryptography and Security
Computer Science - Distributed, Parallel, and Cluster Computing
Computer Science - Learning
Statistics - Machine Learning
title Mitigating Sybils in Federated Learning Poisoning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-28T07%3A00%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Mitigating%20Sybils%20in%20Federated%20Learning%20Poisoning&rft.au=Fung,%20Clement&rft.date=2018-08-14&rft_id=info:doi/10.48550/arxiv.1808.04866&rft_dat=%3Carxiv_GOX%3E1808_04866%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true