Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning

The high cost of communicating gradients is a major bottleneck for federated learning, as the bandwidth of the participating user devices is limited. Existing gradient compression algorithms are mainly designed for data centers with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration c...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Dai, Xinyan, Yan, Xiao, Zhou, Kaiwen, Yang, Han, Ng, Kelvin K. W, Cheng, James, Fan, Yu
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Dai, Xinyan
Yan, Xiao
Zhou, Kaiwen
Yang, Han
Ng, Kelvin K. W
Cheng, James
Fan, Yu
description The high cost of communicating gradients is a major bottleneck for federated learning, as the bandwidth of the participating user devices is limited. Existing gradient compression algorithms are mainly designed for data centers with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration communication cost at best, where $d$ is the size of the model. We propose hyper-sphere quantization (HSQ), a general framework that can be configured to achieve a continuum of trade-offs between communication efficiency and gradient accuracy. In particular, at the high compression ratio end, HSQ provides a low per-iteration communication cost of $O(\log d)$, which is favorable for federated learning. We prove the convergence of HSQ theoretically and show by experiments that HSQ significantly reduces the communication cost of model training without hurting convergence accuracy.
doi_str_mv 10.48550/arxiv.1911.04655
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1911_04655</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1911_04655</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-2ab7c8abda24701196e31a5dfd41b16c7eaab45c3630be1777f3dd100cc086303</originalsourceid><addsrcrecordid>eNotz81OwzAQBGBfOKDCA3DCL-DgreM45YZCf5CCEGrv0dpegyXiRCZFlKcHAqfRzGGkj7ErkEVZay1vMH_GjwJWAIUsK63P2ePuNFIW-_GVMvHnI6YpfuEUh3TLm6Hvjym6uYp1CNFFShPfb-95GDLfkKeME3neEuYU08sFOwv49k6X_7lgh8360OxE-7R9aO5agZXRYonWuBqtx2VpJMCqIgWoffAlWKicIURbaqcqJS2BMSYo70FK52T9s6kFu_67nT3dmGOP-dT9urrZpb4BW19Ikg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning</title><source>arXiv.org</source><creator>Dai, Xinyan ; Yan, Xiao ; Zhou, Kaiwen ; Yang, Han ; Ng, Kelvin K. W ; Cheng, James ; Fan, Yu</creator><creatorcontrib>Dai, Xinyan ; Yan, Xiao ; Zhou, Kaiwen ; Yang, Han ; Ng, Kelvin K. W ; Cheng, James ; Fan, Yu</creatorcontrib><description>The high cost of communicating gradients is a major bottleneck for federated learning, as the bandwidth of the participating user devices is limited. Existing gradient compression algorithms are mainly designed for data centers with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration communication cost at best, where $d$ is the size of the model. We propose hyper-sphere quantization (HSQ), a general framework that can be configured to achieve a continuum of trade-offs between communication efficiency and gradient accuracy. In particular, at the high compression ratio end, HSQ provides a low per-iteration communication cost of $O(\log d)$, which is favorable for federated learning. We prove the convergence of HSQ theoretically and show by experiments that HSQ significantly reduces the communication cost of model training without hurting convergence accuracy.</description><identifier>DOI: 10.48550/arxiv.1911.04655</identifier><language>eng</language><subject>Computer Science - Information Retrieval ; Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2019-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1911.04655$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1911.04655$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Dai, Xinyan</creatorcontrib><creatorcontrib>Yan, Xiao</creatorcontrib><creatorcontrib>Zhou, Kaiwen</creatorcontrib><creatorcontrib>Yang, Han</creatorcontrib><creatorcontrib>Ng, Kelvin K. W</creatorcontrib><creatorcontrib>Cheng, James</creatorcontrib><creatorcontrib>Fan, Yu</creatorcontrib><title>Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning</title><description>The high cost of communicating gradients is a major bottleneck for federated learning, as the bandwidth of the participating user devices is limited. Existing gradient compression algorithms are mainly designed for data centers with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration communication cost at best, where $d$ is the size of the model. We propose hyper-sphere quantization (HSQ), a general framework that can be configured to achieve a continuum of trade-offs between communication efficiency and gradient accuracy. In particular, at the high compression ratio end, HSQ provides a low per-iteration communication cost of $O(\log d)$, which is favorable for federated learning. We prove the convergence of HSQ theoretically and show by experiments that HSQ significantly reduces the communication cost of model training without hurting convergence accuracy.</description><subject>Computer Science - Information Retrieval</subject><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz81OwzAQBGBfOKDCA3DCL-DgreM45YZCf5CCEGrv0dpegyXiRCZFlKcHAqfRzGGkj7ErkEVZay1vMH_GjwJWAIUsK63P2ePuNFIW-_GVMvHnI6YpfuEUh3TLm6Hvjym6uYp1CNFFShPfb-95GDLfkKeME3neEuYU08sFOwv49k6X_7lgh8360OxE-7R9aO5agZXRYonWuBqtx2VpJMCqIgWoffAlWKicIURbaqcqJS2BMSYo70FK52T9s6kFu_67nT3dmGOP-dT9urrZpb4BW19Ikg</recordid><startdate>20191111</startdate><enddate>20191111</enddate><creator>Dai, Xinyan</creator><creator>Yan, Xiao</creator><creator>Zhou, Kaiwen</creator><creator>Yang, Han</creator><creator>Ng, Kelvin K. W</creator><creator>Cheng, James</creator><creator>Fan, Yu</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20191111</creationdate><title>Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning</title><author>Dai, Xinyan ; Yan, Xiao ; Zhou, Kaiwen ; Yang, Han ; Ng, Kelvin K. W ; Cheng, James ; Fan, Yu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-2ab7c8abda24701196e31a5dfd41b16c7eaab45c3630be1777f3dd100cc086303</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Information Retrieval</topic><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Dai, Xinyan</creatorcontrib><creatorcontrib>Yan, Xiao</creatorcontrib><creatorcontrib>Zhou, Kaiwen</creatorcontrib><creatorcontrib>Yang, Han</creatorcontrib><creatorcontrib>Ng, Kelvin K. W</creatorcontrib><creatorcontrib>Cheng, James</creatorcontrib><creatorcontrib>Fan, Yu</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Dai, Xinyan</au><au>Yan, Xiao</au><au>Zhou, Kaiwen</au><au>Yang, Han</au><au>Ng, Kelvin K. W</au><au>Cheng, James</au><au>Fan, Yu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning</atitle><date>2019-11-11</date><risdate>2019</risdate><abstract>The high cost of communicating gradients is a major bottleneck for federated learning, as the bandwidth of the participating user devices is limited. Existing gradient compression algorithms are mainly designed for data centers with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration communication cost at best, where $d$ is the size of the model. We propose hyper-sphere quantization (HSQ), a general framework that can be configured to achieve a continuum of trade-offs between communication efficiency and gradient accuracy. In particular, at the high compression ratio end, HSQ provides a low per-iteration communication cost of $O(\log d)$, which is favorable for federated learning. We prove the convergence of HSQ theoretically and show by experiments that HSQ significantly reduces the communication cost of model training without hurting convergence accuracy.</abstract><doi>10.48550/arxiv.1911.04655</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.1911.04655
ispartof
issn
language eng
recordid cdi_arxiv_primary_1911_04655
source arXiv.org
subjects Computer Science - Information Retrieval
Computer Science - Learning
Statistics - Machine Learning
title Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T10%3A49%3A58IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Hyper-Sphere%20Quantization:%20Communication-Efficient%20SGD%20for%20Federated%20Learning&rft.au=Dai,%20Xinyan&rft.date=2019-11-11&rft_id=info:doi/10.48550/arxiv.1911.04655&rft_dat=%3Carxiv_GOX%3E1911_04655%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true