Federated Learning with Buffered Asynchronous Aggregation

Scalability and privacy are two critical concerns for cross-device federated learning (FL) systems. In this work, we identify that synchronous FL - synchronized aggregation of client updates in FL - cannot scale efficiently beyond a few hundred clients training in parallel. It leads to diminishing r...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Nguyen, John, Malik, Kshitiz, Zhan, Hongyuan, Yousefpour, Ashkan, Rabbat, Michael, Malek, Mani, Huba, Dzmitry
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Nguyen, John
Malik, Kshitiz
Zhan, Hongyuan
Yousefpour, Ashkan
Rabbat, Michael
Malek, Mani
Huba, Dzmitry
description Scalability and privacy are two critical concerns for cross-device federated learning (FL) systems. In this work, we identify that synchronous FL - synchronized aggregation of client updates in FL - cannot scale efficiently beyond a few hundred clients training in parallel. It leads to diminishing returns in model performance and training speed, analogous to large-batch training. On the other hand, asynchronous aggregation of client updates in FL (i.e., asynchronous FL) alleviates the scalability issue. However, aggregating individual client updates is incompatible with Secure Aggregation, which could result in an undesirable level of privacy for the system. To address these concerns, we propose a novel buffered asynchronous aggregation method, FedBuff, that is agnostic to the choice of optimizer, and combines the best properties of synchronous and asynchronous FL. We empirically demonstrate that FedBuff is 3.3x more efficient than synchronous FL and up to 2.5x more efficient than asynchronous FL, while being compatible with privacy-preserving technologies such as Secure Aggregation and differential privacy. We provide theoretical convergence guarantees in a smooth non-convex setting. Finally, we show that under differentially private training, FedBuff can outperform FedAvgM at low privacy settings and achieve the same utility for higher privacy settings.
doi_str_mv 10.48550/arxiv.2106.06639
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2106_06639</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2106_06639</sourcerecordid><originalsourceid>FETCH-LOGICAL-a679-4697ca6067039a12bf612e7268ee7033bd030c7fac9314870164497a04c02ad23</originalsourceid><addsrcrecordid>eNotj71OwzAURr0woJYHYCIvkHD90-t4DBUFpEhduke3znVqiTrISYG-PaUwfdIZPp0jxL2EytSrFTxS_o6flZKAFSBqdyvchnvONHNftEw5xTQUX3E-FE-nEDhfcDOdkz_kMY2nqWiGIfNAcxzTUtwEep_47n8XYrd53q1fy3b78rZu2pLQutKgs54Q0IJ2JNU-oFRsFdbMF6T3PWjwNpB3WpragkRjnCUwHhT1Si_Ew9_t1b37yPFI-dz9NnTXBv0DtPtAPg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Federated Learning with Buffered Asynchronous Aggregation</title><source>arXiv.org</source><creator>Nguyen, John ; Malik, Kshitiz ; Zhan, Hongyuan ; Yousefpour, Ashkan ; Rabbat, Michael ; Malek, Mani ; Huba, Dzmitry</creator><creatorcontrib>Nguyen, John ; Malik, Kshitiz ; Zhan, Hongyuan ; Yousefpour, Ashkan ; Rabbat, Michael ; Malek, Mani ; Huba, Dzmitry</creatorcontrib><description>Scalability and privacy are two critical concerns for cross-device federated learning (FL) systems. In this work, we identify that synchronous FL - synchronized aggregation of client updates in FL - cannot scale efficiently beyond a few hundred clients training in parallel. It leads to diminishing returns in model performance and training speed, analogous to large-batch training. On the other hand, asynchronous aggregation of client updates in FL (i.e., asynchronous FL) alleviates the scalability issue. However, aggregating individual client updates is incompatible with Secure Aggregation, which could result in an undesirable level of privacy for the system. To address these concerns, we propose a novel buffered asynchronous aggregation method, FedBuff, that is agnostic to the choice of optimizer, and combines the best properties of synchronous and asynchronous FL. We empirically demonstrate that FedBuff is 3.3x more efficient than synchronous FL and up to 2.5x more efficient than asynchronous FL, while being compatible with privacy-preserving technologies such as Secure Aggregation and differential privacy. We provide theoretical convergence guarantees in a smooth non-convex setting. Finally, we show that under differentially private training, FedBuff can outperform FedAvgM at low privacy settings and achieve the same utility for higher privacy settings.</description><identifier>DOI: 10.48550/arxiv.2106.06639</identifier><language>eng</language><subject>Computer Science - Learning</subject><creationdate>2021-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,776,881</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2106.06639$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2106.06639$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Nguyen, John</creatorcontrib><creatorcontrib>Malik, Kshitiz</creatorcontrib><creatorcontrib>Zhan, Hongyuan</creatorcontrib><creatorcontrib>Yousefpour, Ashkan</creatorcontrib><creatorcontrib>Rabbat, Michael</creatorcontrib><creatorcontrib>Malek, Mani</creatorcontrib><creatorcontrib>Huba, Dzmitry</creatorcontrib><title>Federated Learning with Buffered Asynchronous Aggregation</title><description>Scalability and privacy are two critical concerns for cross-device federated learning (FL) systems. In this work, we identify that synchronous FL - synchronized aggregation of client updates in FL - cannot scale efficiently beyond a few hundred clients training in parallel. It leads to diminishing returns in model performance and training speed, analogous to large-batch training. On the other hand, asynchronous aggregation of client updates in FL (i.e., asynchronous FL) alleviates the scalability issue. However, aggregating individual client updates is incompatible with Secure Aggregation, which could result in an undesirable level of privacy for the system. To address these concerns, we propose a novel buffered asynchronous aggregation method, FedBuff, that is agnostic to the choice of optimizer, and combines the best properties of synchronous and asynchronous FL. We empirically demonstrate that FedBuff is 3.3x more efficient than synchronous FL and up to 2.5x more efficient than asynchronous FL, while being compatible with privacy-preserving technologies such as Secure Aggregation and differential privacy. We provide theoretical convergence guarantees in a smooth non-convex setting. Finally, we show that under differentially private training, FedBuff can outperform FedAvgM at low privacy settings and achieve the same utility for higher privacy settings.</description><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2021</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotj71OwzAURr0woJYHYCIvkHD90-t4DBUFpEhduke3znVqiTrISYG-PaUwfdIZPp0jxL2EytSrFTxS_o6flZKAFSBqdyvchnvONHNftEw5xTQUX3E-FE-nEDhfcDOdkz_kMY2nqWiGIfNAcxzTUtwEep_47n8XYrd53q1fy3b78rZu2pLQutKgs54Q0IJ2JNU-oFRsFdbMF6T3PWjwNpB3WpragkRjnCUwHhT1Si_Ew9_t1b37yPFI-dz9NnTXBv0DtPtAPg</recordid><startdate>20210611</startdate><enddate>20210611</enddate><creator>Nguyen, John</creator><creator>Malik, Kshitiz</creator><creator>Zhan, Hongyuan</creator><creator>Yousefpour, Ashkan</creator><creator>Rabbat, Michael</creator><creator>Malek, Mani</creator><creator>Huba, Dzmitry</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20210611</creationdate><title>Federated Learning with Buffered Asynchronous Aggregation</title><author>Nguyen, John ; Malik, Kshitiz ; Zhan, Hongyuan ; Yousefpour, Ashkan ; Rabbat, Michael ; Malek, Mani ; Huba, Dzmitry</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a679-4697ca6067039a12bf612e7268ee7033bd030c7fac9314870164497a04c02ad23</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2021</creationdate><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Nguyen, John</creatorcontrib><creatorcontrib>Malik, Kshitiz</creatorcontrib><creatorcontrib>Zhan, Hongyuan</creatorcontrib><creatorcontrib>Yousefpour, Ashkan</creatorcontrib><creatorcontrib>Rabbat, Michael</creatorcontrib><creatorcontrib>Malek, Mani</creatorcontrib><creatorcontrib>Huba, Dzmitry</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Nguyen, John</au><au>Malik, Kshitiz</au><au>Zhan, Hongyuan</au><au>Yousefpour, Ashkan</au><au>Rabbat, Michael</au><au>Malek, Mani</au><au>Huba, Dzmitry</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Federated Learning with Buffered Asynchronous Aggregation</atitle><date>2021-06-11</date><risdate>2021</risdate><abstract>Scalability and privacy are two critical concerns for cross-device federated learning (FL) systems. In this work, we identify that synchronous FL - synchronized aggregation of client updates in FL - cannot scale efficiently beyond a few hundred clients training in parallel. It leads to diminishing returns in model performance and training speed, analogous to large-batch training. On the other hand, asynchronous aggregation of client updates in FL (i.e., asynchronous FL) alleviates the scalability issue. However, aggregating individual client updates is incompatible with Secure Aggregation, which could result in an undesirable level of privacy for the system. To address these concerns, we propose a novel buffered asynchronous aggregation method, FedBuff, that is agnostic to the choice of optimizer, and combines the best properties of synchronous and asynchronous FL. We empirically demonstrate that FedBuff is 3.3x more efficient than synchronous FL and up to 2.5x more efficient than asynchronous FL, while being compatible with privacy-preserving technologies such as Secure Aggregation and differential privacy. We provide theoretical convergence guarantees in a smooth non-convex setting. Finally, we show that under differentially private training, FedBuff can outperform FedAvgM at low privacy settings and achieve the same utility for higher privacy settings.</abstract><doi>10.48550/arxiv.2106.06639</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2106.06639
ispartof
issn
language eng
recordid cdi_arxiv_primary_2106_06639
source arXiv.org
subjects Computer Science - Learning
title Federated Learning with Buffered Asynchronous Aggregation
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-02-05T09%3A06%3A40IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Federated%20Learning%20with%20Buffered%20Asynchronous%20Aggregation&rft.au=Nguyen,%20John&rft.date=2021-06-11&rft_id=info:doi/10.48550/arxiv.2106.06639&rft_dat=%3Carxiv_GOX%3E2106_06639%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true