Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning
The high cost of communicating gradients is a major bottleneck for federated learning, as the bandwidth of the participating user devices is limited. Existing gradient compression algorithms are mainly designed for data centers with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration c...
Gespeichert in:
Hauptverfasser: | , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Dai, Xinyan Yan, Xiao Zhou, Kaiwen Yang, Han Ng, Kelvin K. W Cheng, James Fan, Yu |
description | The high cost of communicating gradients is a major bottleneck for federated
learning, as the bandwidth of the participating user devices is limited.
Existing gradient compression algorithms are mainly designed for data centers
with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration
communication cost at best, where $d$ is the size of the model. We propose
hyper-sphere quantization (HSQ), a general framework that can be configured to
achieve a continuum of trade-offs between communication efficiency and gradient
accuracy. In particular, at the high compression ratio end, HSQ provides a low
per-iteration communication cost of $O(\log d)$, which is favorable for
federated learning. We prove the convergence of HSQ theoretically and show by
experiments that HSQ significantly reduces the communication cost of model
training without hurting convergence accuracy. |
doi_str_mv | 10.48550/arxiv.1911.04655 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_1911_04655</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>1911_04655</sourcerecordid><originalsourceid>FETCH-LOGICAL-a675-2ab7c8abda24701196e31a5dfd41b16c7eaab45c3630be1777f3dd100cc086303</originalsourceid><addsrcrecordid>eNotz81OwzAQBGBfOKDCA3DCL-DgreM45YZCf5CCEGrv0dpegyXiRCZFlKcHAqfRzGGkj7ErkEVZay1vMH_GjwJWAIUsK63P2ePuNFIW-_GVMvHnI6YpfuEUh3TLm6Hvjym6uYp1CNFFShPfb-95GDLfkKeME3neEuYU08sFOwv49k6X_7lgh8360OxE-7R9aO5agZXRYonWuBqtx2VpJMCqIgWoffAlWKicIURbaqcqJS2BMSYo70FK52T9s6kFu_67nT3dmGOP-dT9urrZpb4BW19Ikg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning</title><source>arXiv.org</source><creator>Dai, Xinyan ; Yan, Xiao ; Zhou, Kaiwen ; Yang, Han ; Ng, Kelvin K. W ; Cheng, James ; Fan, Yu</creator><creatorcontrib>Dai, Xinyan ; Yan, Xiao ; Zhou, Kaiwen ; Yang, Han ; Ng, Kelvin K. W ; Cheng, James ; Fan, Yu</creatorcontrib><description>The high cost of communicating gradients is a major bottleneck for federated
learning, as the bandwidth of the participating user devices is limited.
Existing gradient compression algorithms are mainly designed for data centers
with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration
communication cost at best, where $d$ is the size of the model. We propose
hyper-sphere quantization (HSQ), a general framework that can be configured to
achieve a continuum of trade-offs between communication efficiency and gradient
accuracy. In particular, at the high compression ratio end, HSQ provides a low
per-iteration communication cost of $O(\log d)$, which is favorable for
federated learning. We prove the convergence of HSQ theoretically and show by
experiments that HSQ significantly reduces the communication cost of model
training without hurting convergence accuracy.</description><identifier>DOI: 10.48550/arxiv.1911.04655</identifier><language>eng</language><subject>Computer Science - Information Retrieval ; Computer Science - Learning ; Statistics - Machine Learning</subject><creationdate>2019-11</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/1911.04655$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.1911.04655$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Dai, Xinyan</creatorcontrib><creatorcontrib>Yan, Xiao</creatorcontrib><creatorcontrib>Zhou, Kaiwen</creatorcontrib><creatorcontrib>Yang, Han</creatorcontrib><creatorcontrib>Ng, Kelvin K. W</creatorcontrib><creatorcontrib>Cheng, James</creatorcontrib><creatorcontrib>Fan, Yu</creatorcontrib><title>Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning</title><description>The high cost of communicating gradients is a major bottleneck for federated
learning, as the bandwidth of the participating user devices is limited.
Existing gradient compression algorithms are mainly designed for data centers
with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration
communication cost at best, where $d$ is the size of the model. We propose
hyper-sphere quantization (HSQ), a general framework that can be configured to
achieve a continuum of trade-offs between communication efficiency and gradient
accuracy. In particular, at the high compression ratio end, HSQ provides a low
per-iteration communication cost of $O(\log d)$, which is favorable for
federated learning. We prove the convergence of HSQ theoretically and show by
experiments that HSQ significantly reduces the communication cost of model
training without hurting convergence accuracy.</description><subject>Computer Science - Information Retrieval</subject><subject>Computer Science - Learning</subject><subject>Statistics - Machine Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2019</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz81OwzAQBGBfOKDCA3DCL-DgreM45YZCf5CCEGrv0dpegyXiRCZFlKcHAqfRzGGkj7ErkEVZay1vMH_GjwJWAIUsK63P2ePuNFIW-_GVMvHnI6YpfuEUh3TLm6Hvjym6uYp1CNFFShPfb-95GDLfkKeME3neEuYU08sFOwv49k6X_7lgh8360OxE-7R9aO5agZXRYonWuBqtx2VpJMCqIgWoffAlWKicIURbaqcqJS2BMSYo70FK52T9s6kFu_67nT3dmGOP-dT9urrZpb4BW19Ikg</recordid><startdate>20191111</startdate><enddate>20191111</enddate><creator>Dai, Xinyan</creator><creator>Yan, Xiao</creator><creator>Zhou, Kaiwen</creator><creator>Yang, Han</creator><creator>Ng, Kelvin K. W</creator><creator>Cheng, James</creator><creator>Fan, Yu</creator><scope>AKY</scope><scope>EPD</scope><scope>GOX</scope></search><sort><creationdate>20191111</creationdate><title>Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning</title><author>Dai, Xinyan ; Yan, Xiao ; Zhou, Kaiwen ; Yang, Han ; Ng, Kelvin K. W ; Cheng, James ; Fan, Yu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a675-2ab7c8abda24701196e31a5dfd41b16c7eaab45c3630be1777f3dd100cc086303</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2019</creationdate><topic>Computer Science - Information Retrieval</topic><topic>Computer Science - Learning</topic><topic>Statistics - Machine Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Dai, Xinyan</creatorcontrib><creatorcontrib>Yan, Xiao</creatorcontrib><creatorcontrib>Zhou, Kaiwen</creatorcontrib><creatorcontrib>Yang, Han</creatorcontrib><creatorcontrib>Ng, Kelvin K. W</creatorcontrib><creatorcontrib>Cheng, James</creatorcontrib><creatorcontrib>Fan, Yu</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv Statistics</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Dai, Xinyan</au><au>Yan, Xiao</au><au>Zhou, Kaiwen</au><au>Yang, Han</au><au>Ng, Kelvin K. W</au><au>Cheng, James</au><au>Fan, Yu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning</atitle><date>2019-11-11</date><risdate>2019</risdate><abstract>The high cost of communicating gradients is a major bottleneck for federated
learning, as the bandwidth of the participating user devices is limited.
Existing gradient compression algorithms are mainly designed for data centers
with high-speed network and achieve $O(\sqrt{d} \log d)$ per-iteration
communication cost at best, where $d$ is the size of the model. We propose
hyper-sphere quantization (HSQ), a general framework that can be configured to
achieve a continuum of trade-offs between communication efficiency and gradient
accuracy. In particular, at the high compression ratio end, HSQ provides a low
per-iteration communication cost of $O(\log d)$, which is favorable for
federated learning. We prove the convergence of HSQ theoretically and show by
experiments that HSQ significantly reduces the communication cost of model
training without hurting convergence accuracy.</abstract><doi>10.48550/arxiv.1911.04655</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.1911.04655 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_1911_04655 |
source | arXiv.org |
subjects | Computer Science - Information Retrieval Computer Science - Learning Statistics - Machine Learning |
title | Hyper-Sphere Quantization: Communication-Efficient SGD for Federated Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-27T10%3A49%3A58IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Hyper-Sphere%20Quantization:%20Communication-Efficient%20SGD%20for%20Federated%20Learning&rft.au=Dai,%20Xinyan&rft.date=2019-11-11&rft_id=info:doi/10.48550/arxiv.1911.04655&rft_dat=%3Carxiv_GOX%3E1911_04655%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |