Straggler-Resilient Personalized Federated Learning
Federated Learning is an emerging learning paradigm that allows training models from samples distributed across a large network of clients while respecting privacy and communication restrictions. Despite its success, federated learning faces several challenges related to its decentralized nature. In...
Gespeichert in:
Hauptverfasser: | , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Tziotis, Isidoros Shen, Zebang Pedarsani, Ramtin Hassani, Hamed Mokhtari, Aryan |
description | Federated Learning is an emerging learning paradigm that allows training
models from samples distributed across a large network of clients while
respecting privacy and communication restrictions. Despite its success,
federated learning faces several challenges related to its decentralized
nature. In this work, we develop a novel algorithmic procedure with theoretical
speedup guarantees that simultaneously handles two of these hurdles, namely (i)
data heterogeneity, i.e., data distributions can vary substantially across
clients, and (ii) system heterogeneity, i.e., the computational power of the
clients could differ significantly. Our method relies on ideas from
representation learning theory to find a global common representation using all
clients' data and learn a user-specific set of parameters leading to a
personalized solution for each client. Furthermore, our method mitigates the
effects of stragglers by adaptively selecting clients based on their
computational characteristics and statistical significance, thus achieving, for
the first time, near optimal sample complexity and provable logarithmic
speedup. Experimental results support our theoretical findings showing the
superiority of our method over alternative personalized federated schemes in
system and data heterogeneous environments. |
doi_str_mv | 10.48550/arxiv.2206.02078 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2206_02078</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2206_02078</sourcerecordid><originalsourceid>FETCH-LOGICAL-a678-2bf9c157d4e89e04d85986d149de5552a23a97a03b43af942cdeace07d0a6ada3</originalsourceid><addsrcrecordid>eNotzs1qwkAUhuHZuCjqBXSlN5B4Mj-ZmaWIViHQUt2HY85JGIhRJkGsV2-rXX3v6uMR4j2DVDtjYIHxFq6plJCnIMG6N6H2Q8SmaTkm39yHNnA3zL849ucO23Bnmm-YOOLwWwVj7ELXTMSoxrbn6f-OxWGzPqy2SfH5sVstiwRz6xJ5rH2VGUuanWfQ5Ix3OWXaExtjJEqF3iKoo1ZYey0rYqwYLAHmSKjGYva6farLSwwnjD_ln7586tUDXlY_0g</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Straggler-Resilient Personalized Federated Learning</title><source>arXiv.org</source><creator>Tziotis, Isidoros ; Shen, Zebang ; Pedarsani, Ramtin ; Hassani, Hamed ; Mokhtari, Aryan</creator><creatorcontrib>Tziotis, Isidoros ; Shen, Zebang ; Pedarsani, Ramtin ; Hassani, Hamed ; Mokhtari, Aryan</creatorcontrib><description>Federated Learning is an emerging learning paradigm that allows training
models from samples distributed across a large network of clients while
respecting privacy and communication restrictions. Despite its success,
federated learning faces several challenges related to its decentralized
nature. In this work, we develop a novel algorithmic procedure with theoretical
speedup guarantees that simultaneously handles two of these hurdles, namely (i)
data heterogeneity, i.e., data distributions can vary substantially across
clients, and (ii) system heterogeneity, i.e., the computational power of the
clients could differ significantly. Our method relies on ideas from
representation learning theory to find a global common representation using all
clients' data and learn a user-specific set of parameters leading to a
personalized solution for each client. Furthermore, our method mitigates the
effects of stragglers by adaptively selecting clients based on their
computational characteristics and statistical significance, thus achieving, for
the first time, near optimal sample complexity and provable logarithmic
speedup. Experimental results support our theoretical findings showing the
superiority of our method over alternative personalized federated schemes in
system and data heterogeneous environments.</description><identifier>DOI: 10.48550/arxiv.2206.02078</identifier><language>eng</language><subject>Computer Science - Distributed, Parallel, and Cluster Computing ; Computer Science - Learning</subject><creationdate>2022-06</creationdate><rights>http://arxiv.org/licenses/nonexclusive-distrib/1.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2206.02078$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2206.02078$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Tziotis, Isidoros</creatorcontrib><creatorcontrib>Shen, Zebang</creatorcontrib><creatorcontrib>Pedarsani, Ramtin</creatorcontrib><creatorcontrib>Hassani, Hamed</creatorcontrib><creatorcontrib>Mokhtari, Aryan</creatorcontrib><title>Straggler-Resilient Personalized Federated Learning</title><description>Federated Learning is an emerging learning paradigm that allows training
models from samples distributed across a large network of clients while
respecting privacy and communication restrictions. Despite its success,
federated learning faces several challenges related to its decentralized
nature. In this work, we develop a novel algorithmic procedure with theoretical
speedup guarantees that simultaneously handles two of these hurdles, namely (i)
data heterogeneity, i.e., data distributions can vary substantially across
clients, and (ii) system heterogeneity, i.e., the computational power of the
clients could differ significantly. Our method relies on ideas from
representation learning theory to find a global common representation using all
clients' data and learn a user-specific set of parameters leading to a
personalized solution for each client. Furthermore, our method mitigates the
effects of stragglers by adaptively selecting clients based on their
computational characteristics and statistical significance, thus achieving, for
the first time, near optimal sample complexity and provable logarithmic
speedup. Experimental results support our theoretical findings showing the
superiority of our method over alternative personalized federated schemes in
system and data heterogeneous environments.</description><subject>Computer Science - Distributed, Parallel, and Cluster Computing</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotzs1qwkAUhuHZuCjqBXSlN5B4Mj-ZmaWIViHQUt2HY85JGIhRJkGsV2-rXX3v6uMR4j2DVDtjYIHxFq6plJCnIMG6N6H2Q8SmaTkm39yHNnA3zL849ucO23Bnmm-YOOLwWwVj7ELXTMSoxrbn6f-OxWGzPqy2SfH5sVstiwRz6xJ5rH2VGUuanWfQ5Ix3OWXaExtjJEqF3iKoo1ZYey0rYqwYLAHmSKjGYva6farLSwwnjD_ln7586tUDXlY_0g</recordid><startdate>20220604</startdate><enddate>20220604</enddate><creator>Tziotis, Isidoros</creator><creator>Shen, Zebang</creator><creator>Pedarsani, Ramtin</creator><creator>Hassani, Hamed</creator><creator>Mokhtari, Aryan</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20220604</creationdate><title>Straggler-Resilient Personalized Federated Learning</title><author>Tziotis, Isidoros ; Shen, Zebang ; Pedarsani, Ramtin ; Hassani, Hamed ; Mokhtari, Aryan</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a678-2bf9c157d4e89e04d85986d149de5552a23a97a03b43af942cdeace07d0a6ada3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Distributed, Parallel, and Cluster Computing</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Tziotis, Isidoros</creatorcontrib><creatorcontrib>Shen, Zebang</creatorcontrib><creatorcontrib>Pedarsani, Ramtin</creatorcontrib><creatorcontrib>Hassani, Hamed</creatorcontrib><creatorcontrib>Mokhtari, Aryan</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Tziotis, Isidoros</au><au>Shen, Zebang</au><au>Pedarsani, Ramtin</au><au>Hassani, Hamed</au><au>Mokhtari, Aryan</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Straggler-Resilient Personalized Federated Learning</atitle><date>2022-06-04</date><risdate>2022</risdate><abstract>Federated Learning is an emerging learning paradigm that allows training
models from samples distributed across a large network of clients while
respecting privacy and communication restrictions. Despite its success,
federated learning faces several challenges related to its decentralized
nature. In this work, we develop a novel algorithmic procedure with theoretical
speedup guarantees that simultaneously handles two of these hurdles, namely (i)
data heterogeneity, i.e., data distributions can vary substantially across
clients, and (ii) system heterogeneity, i.e., the computational power of the
clients could differ significantly. Our method relies on ideas from
representation learning theory to find a global common representation using all
clients' data and learn a user-specific set of parameters leading to a
personalized solution for each client. Furthermore, our method mitigates the
effects of stragglers by adaptively selecting clients based on their
computational characteristics and statistical significance, thus achieving, for
the first time, near optimal sample complexity and provable logarithmic
speedup. Experimental results support our theoretical findings showing the
superiority of our method over alternative personalized federated schemes in
system and data heterogeneous environments.</abstract><doi>10.48550/arxiv.2206.02078</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2206.02078 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2206_02078 |
source | arXiv.org |
subjects | Computer Science - Distributed, Parallel, and Cluster Computing Computer Science - Learning |
title | Straggler-Resilient Personalized Federated Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-11T00%3A43%3A05IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Straggler-Resilient%20Personalized%20Federated%20Learning&rft.au=Tziotis,%20Isidoros&rft.date=2022-06-04&rft_id=info:doi/10.48550/arxiv.2206.02078&rft_dat=%3Carxiv_GOX%3E2206_02078%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |