A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg
Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenge...
Gespeichert in:
Veröffentlicht in: | The Journal of artificial intelligence research 2023-01, Vol.78, p.1143-1200 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1200 |
---|---|
container_issue | |
container_start_page | 1143 |
container_title | The Journal of artificial intelligence research |
container_volume | 78 |
creator | Qu, Zhaonan Lin, Kaixiang Li, Zhaojian Zhou, Jiayu Zhou, Zhengyuan |
description | Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenges in understanding the convergence of FL algorithms, particularly regarding how convergence scales with the number of participating devices. In this paper, we focus on Federated Averaging (FedAvg), one of the most popular and effective FL algorithms in use today, as well as its Nesterov accelerated variant, and conduct a systematic study of how their convergence scale with the number of participating devices under non-i.i.d. data and partial participation in convex settings. We provide a unified analysis that establishes convergence guarantees for FedAvg under strongly convex, convex, and overparameterized strongly convex problems. We show that FedAvg enjoys linear speedup in each case, although with different convergence rates and communication efficiencies. For strongly convex and convex problems, we also characterize the corresponding convergence rates for the Nesterov accelerated FedAvg algorithm, which are the first linear speedup guarantees for momentum variants of FedAvg in convex settings. Empirical studies of the algorithms in various settings have supported our theoretical results. |
doi_str_mv | 10.1613/jair.1.15180 |
format | Article |
fullrecord | <record><control><sourceid>crossref</sourceid><recordid>TN_cdi_crossref_primary_10_1613_jair_1_15180</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>10_1613_jair_1_15180</sourcerecordid><originalsourceid>FETCH-LOGICAL-c192t-fae3cbe5ba84b3680bb3a76759405e9e0fe12171cc9fd16e419cc013fbf2a5013</originalsourceid><addsrcrecordid>eNpNkLFOwzAURS0EEqWw8QH-AFL84jiJx6iiFCmCATpbtvMcuSpJZIdI_XsSYGC6Zzi6wyHkHtgGcuCPR-3DBjYgoGQXZAWsyBNZiOLyH1-TmxiPjIHM0nJF6ooeOu88NrT2HepA3wfE5mugVadP5-gj7R3dYYNBj7NUTTO0vmup7hr6inHE0E-LUE3tLbly-hTx7m_X5LB7-tjuk_rt-WVb1YkFmY6J08itQWF0mRmel8wYrou8EDJjAiUyh5BCAdZK10COGUhrGXBnXKrFDGvy8PtrQx9jQKeG4D91OCtgaimhlhIK1E8J_g1PGFHm</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg</title><source>DOAJ Directory of Open Access Journals</source><source>EZB-FREE-00999 freely available EZB journals</source><source>Free E- Journals</source><creator>Qu, Zhaonan ; Lin, Kaixiang ; Li, Zhaojian ; Zhou, Jiayu ; Zhou, Zhengyuan</creator><creatorcontrib>Qu, Zhaonan ; Lin, Kaixiang ; Li, Zhaojian ; Zhou, Jiayu ; Zhou, Zhengyuan</creatorcontrib><description>Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenges in understanding the convergence of FL algorithms, particularly regarding how convergence scales with the number of participating devices. In this paper, we focus on Federated Averaging (FedAvg), one of the most popular and effective FL algorithms in use today, as well as its Nesterov accelerated variant, and conduct a systematic study of how their convergence scale with the number of participating devices under non-i.i.d. data and partial participation in convex settings. We provide a unified analysis that establishes convergence guarantees for FedAvg under strongly convex, convex, and overparameterized strongly convex problems. We show that FedAvg enjoys linear speedup in each case, although with different convergence rates and communication efficiencies. For strongly convex and convex problems, we also characterize the corresponding convergence rates for the Nesterov accelerated FedAvg algorithm, which are the first linear speedup guarantees for momentum variants of FedAvg in convex settings. Empirical studies of the algorithms in various settings have supported our theoretical results.</description><identifier>ISSN: 1076-9757</identifier><identifier>EISSN: 1076-9757</identifier><identifier>DOI: 10.1613/jair.1.15180</identifier><language>eng</language><ispartof>The Journal of artificial intelligence research, 2023-01, Vol.78, p.1143-1200</ispartof><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,864,27924,27925</link.rule.ids></links><search><creatorcontrib>Qu, Zhaonan</creatorcontrib><creatorcontrib>Lin, Kaixiang</creatorcontrib><creatorcontrib>Li, Zhaojian</creatorcontrib><creatorcontrib>Zhou, Jiayu</creatorcontrib><creatorcontrib>Zhou, Zhengyuan</creatorcontrib><title>A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg</title><title>The Journal of artificial intelligence research</title><description>Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenges in understanding the convergence of FL algorithms, particularly regarding how convergence scales with the number of participating devices. In this paper, we focus on Federated Averaging (FedAvg), one of the most popular and effective FL algorithms in use today, as well as its Nesterov accelerated variant, and conduct a systematic study of how their convergence scale with the number of participating devices under non-i.i.d. data and partial participation in convex settings. We provide a unified analysis that establishes convergence guarantees for FedAvg under strongly convex, convex, and overparameterized strongly convex problems. We show that FedAvg enjoys linear speedup in each case, although with different convergence rates and communication efficiencies. For strongly convex and convex problems, we also characterize the corresponding convergence rates for the Nesterov accelerated FedAvg algorithm, which are the first linear speedup guarantees for momentum variants of FedAvg in convex settings. Empirical studies of the algorithms in various settings have supported our theoretical results.</description><issn>1076-9757</issn><issn>1076-9757</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpNkLFOwzAURS0EEqWw8QH-AFL84jiJx6iiFCmCATpbtvMcuSpJZIdI_XsSYGC6Zzi6wyHkHtgGcuCPR-3DBjYgoGQXZAWsyBNZiOLyH1-TmxiPjIHM0nJF6ooeOu88NrT2HepA3wfE5mugVadP5-gj7R3dYYNBj7NUTTO0vmup7hr6inHE0E-LUE3tLbly-hTx7m_X5LB7-tjuk_rt-WVb1YkFmY6J08itQWF0mRmel8wYrou8EDJjAiUyh5BCAdZK10COGUhrGXBnXKrFDGvy8PtrQx9jQKeG4D91OCtgaimhlhIK1E8J_g1PGFHm</recordid><startdate>20230101</startdate><enddate>20230101</enddate><creator>Qu, Zhaonan</creator><creator>Lin, Kaixiang</creator><creator>Li, Zhaojian</creator><creator>Zhou, Jiayu</creator><creator>Zhou, Zhengyuan</creator><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20230101</creationdate><title>A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg</title><author>Qu, Zhaonan ; Lin, Kaixiang ; Li, Zhaojian ; Zhou, Jiayu ; Zhou, Zhengyuan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c192t-fae3cbe5ba84b3680bb3a76759405e9e0fe12171cc9fd16e419cc013fbf2a5013</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Qu, Zhaonan</creatorcontrib><creatorcontrib>Lin, Kaixiang</creatorcontrib><creatorcontrib>Li, Zhaojian</creatorcontrib><creatorcontrib>Zhou, Jiayu</creatorcontrib><creatorcontrib>Zhou, Zhengyuan</creatorcontrib><collection>CrossRef</collection><jtitle>The Journal of artificial intelligence research</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Qu, Zhaonan</au><au>Lin, Kaixiang</au><au>Li, Zhaojian</au><au>Zhou, Jiayu</au><au>Zhou, Zhengyuan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg</atitle><jtitle>The Journal of artificial intelligence research</jtitle><date>2023-01-01</date><risdate>2023</risdate><volume>78</volume><spage>1143</spage><epage>1200</epage><pages>1143-1200</pages><issn>1076-9757</issn><eissn>1076-9757</eissn><abstract>Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenges in understanding the convergence of FL algorithms, particularly regarding how convergence scales with the number of participating devices. In this paper, we focus on Federated Averaging (FedAvg), one of the most popular and effective FL algorithms in use today, as well as its Nesterov accelerated variant, and conduct a systematic study of how their convergence scale with the number of participating devices under non-i.i.d. data and partial participation in convex settings. We provide a unified analysis that establishes convergence guarantees for FedAvg under strongly convex, convex, and overparameterized strongly convex problems. We show that FedAvg enjoys linear speedup in each case, although with different convergence rates and communication efficiencies. For strongly convex and convex problems, we also characterize the corresponding convergence rates for the Nesterov accelerated FedAvg algorithm, which are the first linear speedup guarantees for momentum variants of FedAvg in convex settings. Empirical studies of the algorithms in various settings have supported our theoretical results.</abstract><doi>10.1613/jair.1.15180</doi><tpages>58</tpages></addata></record> |
fulltext | fulltext |
identifier | ISSN: 1076-9757 |
ispartof | The Journal of artificial intelligence research, 2023-01, Vol.78, p.1143-1200 |
issn | 1076-9757 1076-9757 |
language | eng |
recordid | cdi_crossref_primary_10_1613_jair_1_15180 |
source | DOAJ Directory of Open Access Journals; EZB-FREE-00999 freely available EZB journals; Free E- Journals |
title | A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T23%3A55%3A10IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-crossref&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Unified%20Linear%20Speedup%20Analysis%20of%20Federated%20Averaging%20and%20Nesterov%20FedAvg&rft.jtitle=The%20Journal%20of%20artificial%20intelligence%20research&rft.au=Qu,%20Zhaonan&rft.date=2023-01-01&rft.volume=78&rft.spage=1143&rft.epage=1200&rft.pages=1143-1200&rft.issn=1076-9757&rft.eissn=1076-9757&rft_id=info:doi/10.1613/jair.1.15180&rft_dat=%3Ccrossref%3E10_1613_jair_1_15180%3C/crossref%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |