A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg

Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenge...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:The Journal of artificial intelligence research 2023-01, Vol.78, p.1143-1200
Hauptverfasser: Qu, Zhaonan, Lin, Kaixiang, Li, Zhaojian, Zhou, Jiayu, Zhou, Zhengyuan
Format: Artikel
Sprache:eng
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1200
container_issue
container_start_page 1143
container_title The Journal of artificial intelligence research
container_volume 78
creator Qu, Zhaonan
Lin, Kaixiang
Li, Zhaojian
Zhou, Jiayu
Zhou, Zhengyuan
description Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenges in understanding the convergence of FL algorithms, particularly regarding how convergence scales with the number of participating devices. In this paper, we focus on Federated Averaging (FedAvg), one of the most popular and effective FL algorithms in use today, as well as its Nesterov accelerated variant, and conduct a systematic study of how their convergence scale with the number of participating devices under non-i.i.d. data and partial participation in convex settings. We provide a unified analysis that establishes convergence guarantees for FedAvg under strongly convex, convex, and overparameterized strongly convex problems. We show that FedAvg enjoys linear speedup in each case, although with different convergence rates and communication efficiencies. For strongly convex and convex problems, we also characterize the corresponding convergence rates for the Nesterov accelerated FedAvg algorithm, which are the first linear speedup guarantees for momentum variants of FedAvg in convex settings. Empirical studies of the algorithms in various settings have supported our theoretical results.
doi_str_mv 10.1613/jair.1.15180
format Article
fullrecord <record><control><sourceid>crossref</sourceid><recordid>TN_cdi_crossref_primary_10_1613_jair_1_15180</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>10_1613_jair_1_15180</sourcerecordid><originalsourceid>FETCH-LOGICAL-c192t-fae3cbe5ba84b3680bb3a76759405e9e0fe12171cc9fd16e419cc013fbf2a5013</originalsourceid><addsrcrecordid>eNpNkLFOwzAURS0EEqWw8QH-AFL84jiJx6iiFCmCATpbtvMcuSpJZIdI_XsSYGC6Zzi6wyHkHtgGcuCPR-3DBjYgoGQXZAWsyBNZiOLyH1-TmxiPjIHM0nJF6ooeOu88NrT2HepA3wfE5mugVadP5-gj7R3dYYNBj7NUTTO0vmup7hr6inHE0E-LUE3tLbly-hTx7m_X5LB7-tjuk_rt-WVb1YkFmY6J08itQWF0mRmel8wYrou8EDJjAiUyh5BCAdZK10COGUhrGXBnXKrFDGvy8PtrQx9jQKeG4D91OCtgaimhlhIK1E8J_g1PGFHm</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg</title><source>DOAJ Directory of Open Access Journals</source><source>EZB-FREE-00999 freely available EZB journals</source><source>Free E- Journals</source><creator>Qu, Zhaonan ; Lin, Kaixiang ; Li, Zhaojian ; Zhou, Jiayu ; Zhou, Zhengyuan</creator><creatorcontrib>Qu, Zhaonan ; Lin, Kaixiang ; Li, Zhaojian ; Zhou, Jiayu ; Zhou, Zhengyuan</creatorcontrib><description>Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenges in understanding the convergence of FL algorithms, particularly regarding how convergence scales with the number of participating devices. In this paper, we focus on Federated Averaging (FedAvg), one of the most popular and effective FL algorithms in use today, as well as its Nesterov accelerated variant, and conduct a systematic study of how their convergence scale with the number of participating devices under non-i.i.d. data and partial participation in convex settings. We provide a unified analysis that establishes convergence guarantees for FedAvg under strongly convex, convex, and overparameterized strongly convex problems. We show that FedAvg enjoys linear speedup in each case, although with different convergence rates and communication efficiencies. For strongly convex and convex problems, we also characterize the corresponding convergence rates for the Nesterov accelerated FedAvg algorithm, which are the first linear speedup guarantees for momentum variants of FedAvg in convex settings. Empirical studies of the algorithms in various settings have supported our theoretical results.</description><identifier>ISSN: 1076-9757</identifier><identifier>EISSN: 1076-9757</identifier><identifier>DOI: 10.1613/jair.1.15180</identifier><language>eng</language><ispartof>The Journal of artificial intelligence research, 2023-01, Vol.78, p.1143-1200</ispartof><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,780,784,864,27924,27925</link.rule.ids></links><search><creatorcontrib>Qu, Zhaonan</creatorcontrib><creatorcontrib>Lin, Kaixiang</creatorcontrib><creatorcontrib>Li, Zhaojian</creatorcontrib><creatorcontrib>Zhou, Jiayu</creatorcontrib><creatorcontrib>Zhou, Zhengyuan</creatorcontrib><title>A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg</title><title>The Journal of artificial intelligence research</title><description>Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenges in understanding the convergence of FL algorithms, particularly regarding how convergence scales with the number of participating devices. In this paper, we focus on Federated Averaging (FedAvg), one of the most popular and effective FL algorithms in use today, as well as its Nesterov accelerated variant, and conduct a systematic study of how their convergence scale with the number of participating devices under non-i.i.d. data and partial participation in convex settings. We provide a unified analysis that establishes convergence guarantees for FedAvg under strongly convex, convex, and overparameterized strongly convex problems. We show that FedAvg enjoys linear speedup in each case, although with different convergence rates and communication efficiencies. For strongly convex and convex problems, we also characterize the corresponding convergence rates for the Nesterov accelerated FedAvg algorithm, which are the first linear speedup guarantees for momentum variants of FedAvg in convex settings. Empirical studies of the algorithms in various settings have supported our theoretical results.</description><issn>1076-9757</issn><issn>1076-9757</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpNkLFOwzAURS0EEqWw8QH-AFL84jiJx6iiFCmCATpbtvMcuSpJZIdI_XsSYGC6Zzi6wyHkHtgGcuCPR-3DBjYgoGQXZAWsyBNZiOLyH1-TmxiPjIHM0nJF6ooeOu88NrT2HepA3wfE5mugVadP5-gj7R3dYYNBj7NUTTO0vmup7hr6inHE0E-LUE3tLbly-hTx7m_X5LB7-tjuk_rt-WVb1YkFmY6J08itQWF0mRmel8wYrou8EDJjAiUyh5BCAdZK10COGUhrGXBnXKrFDGvy8PtrQx9jQKeG4D91OCtgaimhlhIK1E8J_g1PGFHm</recordid><startdate>20230101</startdate><enddate>20230101</enddate><creator>Qu, Zhaonan</creator><creator>Lin, Kaixiang</creator><creator>Li, Zhaojian</creator><creator>Zhou, Jiayu</creator><creator>Zhou, Zhengyuan</creator><scope>AAYXX</scope><scope>CITATION</scope></search><sort><creationdate>20230101</creationdate><title>A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg</title><author>Qu, Zhaonan ; Lin, Kaixiang ; Li, Zhaojian ; Zhou, Jiayu ; Zhou, Zhengyuan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c192t-fae3cbe5ba84b3680bb3a76759405e9e0fe12171cc9fd16e419cc013fbf2a5013</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Qu, Zhaonan</creatorcontrib><creatorcontrib>Lin, Kaixiang</creatorcontrib><creatorcontrib>Li, Zhaojian</creatorcontrib><creatorcontrib>Zhou, Jiayu</creatorcontrib><creatorcontrib>Zhou, Zhengyuan</creatorcontrib><collection>CrossRef</collection><jtitle>The Journal of artificial intelligence research</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Qu, Zhaonan</au><au>Lin, Kaixiang</au><au>Li, Zhaojian</au><au>Zhou, Jiayu</au><au>Zhou, Zhengyuan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg</atitle><jtitle>The Journal of artificial intelligence research</jtitle><date>2023-01-01</date><risdate>2023</risdate><volume>78</volume><spage>1143</spage><epage>1200</epage><pages>1143-1200</pages><issn>1076-9757</issn><eissn>1076-9757</eissn><abstract>Federated learning (FL) learns a model jointly from a set of participating devices without sharing each other’s privately held data. The characteristics of non-i.i.d. data across the network, low device participation, high communication costs, and the mandate that data remain private bring challenges in understanding the convergence of FL algorithms, particularly regarding how convergence scales with the number of participating devices. In this paper, we focus on Federated Averaging (FedAvg), one of the most popular and effective FL algorithms in use today, as well as its Nesterov accelerated variant, and conduct a systematic study of how their convergence scale with the number of participating devices under non-i.i.d. data and partial participation in convex settings. We provide a unified analysis that establishes convergence guarantees for FedAvg under strongly convex, convex, and overparameterized strongly convex problems. We show that FedAvg enjoys linear speedup in each case, although with different convergence rates and communication efficiencies. For strongly convex and convex problems, we also characterize the corresponding convergence rates for the Nesterov accelerated FedAvg algorithm, which are the first linear speedup guarantees for momentum variants of FedAvg in convex settings. Empirical studies of the algorithms in various settings have supported our theoretical results.</abstract><doi>10.1613/jair.1.15180</doi><tpages>58</tpages></addata></record>
fulltext fulltext
identifier ISSN: 1076-9757
ispartof The Journal of artificial intelligence research, 2023-01, Vol.78, p.1143-1200
issn 1076-9757
1076-9757
language eng
recordid cdi_crossref_primary_10_1613_jair_1_15180
source DOAJ Directory of Open Access Journals; EZB-FREE-00999 freely available EZB journals; Free E- Journals
title A Unified Linear Speedup Analysis of Federated Averaging and Nesterov FedAvg
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-26T23%3A55%3A10IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-crossref&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=A%20Unified%20Linear%20Speedup%20Analysis%20of%20Federated%20Averaging%20and%20Nesterov%20FedAvg&rft.jtitle=The%20Journal%20of%20artificial%20intelligence%20research&rft.au=Qu,%20Zhaonan&rft.date=2023-01-01&rft.volume=78&rft.spage=1143&rft.epage=1200&rft.pages=1143-1200&rft.issn=1076-9757&rft.eissn=1076-9757&rft_id=info:doi/10.1613/jair.1.15180&rft_dat=%3Ccrossref%3E10_1613_jair_1_15180%3C/crossref%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true