FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning

In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-bas...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE transaction on neural networks and learning systems 2024-09, Vol.PP, p.1-15
Hauptverfasser: Gao, Dehong, Song, Duanxiao, Shen, Guangyuan, Cai, Xiaoyan, Yang, Libin, Liu, Gongshen, Li, Xiaoyong, Wang, Zhen
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 15
container_issue
container_start_page 1
container_title IEEE transaction on neural networks and learning systems
container_volume PP
creator Gao, Dehong
Song, Duanxiao
Shen, Guangyuan
Cai, Xiaoyan
Yang, Libin
Liu, Gongshen
Li, Xiaoyong
Wang, Zhen
description In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-based client selection constitutes a simple yet efficacious approach to the identification of those clients possessing representative gradient information. Despite the extensive body of research on modified selection methodologies, the majority of prior work is predicated upon the assumption of consistently effective clustering. However, raw gradient-based clustering methods are subject to several challenges: 1) poor effectiveness, the raw high-dimensional gradient of a client is too complex to serve as an appropriate feature for grouping, resulting in large intra-cluster distances and 2) fluctuating effectiveness, due to inherent limitations in clustering, the effectiveness can vary significantly, leading to clusters with diverse levels of heterogeneity. In practice, suboptimal and inconsistent clustering effects can result in clusters with low intra-cluster similarity among clients. The selection of clients from such clusters may impede the overall convergence of training. In this article, we propose, a novel client selection scheme to accelerate the FL convergence by variance reduction. The main idea of is to stratify a compressed model update in order to ensure an excellent grouping effect, and at the same time reduce the cross-client variance by re-allocating the sample chance among different groups based on their diverse heterogeneity. It strikes this convergence acceleration by paying more attention to those client groups with relatively low similarity and then improving the representativeness of the selected subset as much as possible. Theoretically, we demonstrate the critical improvement of the proposed scheme in variance reduction and present equivalence conditions among different client selection methods. We also present the tighter convergence guarantee of the proposed method thanks to the variance reduction. Experimental results confirm the exceeded efficiency of our approach compared to alternatives.
doi_str_mv 10.1109/TNNLS.2024.3438843
format Article
fullrecord <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_miscellaneous_3109420212</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10689614</ieee_id><sourcerecordid>3109420212</sourcerecordid><originalsourceid>FETCH-LOGICAL-c205t-ad46a44487d62bbfb64ad5224fad4978b48fc33b2defd70e30ba2cafeeaa53733</originalsourceid><addsrcrecordid>eNpNkF1LwzAUhoMobsz9ARHJpTedaZKmqXdjWBXKvOiE3ZW0PZFoP2bSIfv3Zm4Oz03OIc_7XjwIXYdkFoYkuV8tl1k-o4TyGeNMSs7O0JiGggbUX-enPV6P0NS5D-JHkEjw5BKNWMJCwaUco3UKdb7KH_Ac54NVg9EGarxoDHQDzqGBajB9h1OrWvju7SfWvcWLvnPGDR5pdjhVbsC-BXzaRzNQtjPd-xW60KpxMD2-E_SWPq4Wz0H2-vSymGdBRUk0BKrmQnHOZVwLWpa6FFzVEaVc-58kliWXumKspDXoOibASKlopTSAUhGLGZugu0PvxvZfW3BD0RpXQdOoDvqtK5iXxb2mkHqUHtDK9s5Z0MXGmlbZXRGSYi-1-JVa7KUWR6k-dHvs35Yt1KfIn0IP3BwAAwD_GoVMRMjZDwj2fDQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3109420212</pqid></control><display><type>article</type><title>FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning</title><source>IEEE Electronic Library (IEL)</source><creator>Gao, Dehong ; Song, Duanxiao ; Shen, Guangyuan ; Cai, Xiaoyan ; Yang, Libin ; Liu, Gongshen ; Li, Xiaoyong ; Wang, Zhen</creator><creatorcontrib>Gao, Dehong ; Song, Duanxiao ; Shen, Guangyuan ; Cai, Xiaoyan ; Yang, Libin ; Liu, Gongshen ; Li, Xiaoyong ; Wang, Zhen</creatorcontrib><description>In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-based client selection constitutes a simple yet efficacious approach to the identification of those clients possessing representative gradient information. Despite the extensive body of research on modified selection methodologies, the majority of prior work is predicated upon the assumption of consistently effective clustering. However, raw gradient-based clustering methods are subject to several challenges: 1) poor effectiveness, the raw high-dimensional gradient of a client is too complex to serve as an appropriate feature for grouping, resulting in large intra-cluster distances and 2) fluctuating effectiveness, due to inherent limitations in clustering, the effectiveness can vary significantly, leading to clusters with diverse levels of heterogeneity. In practice, suboptimal and inconsistent clustering effects can result in clusters with low intra-cluster similarity among clients. The selection of clients from such clusters may impede the overall convergence of training. In this article, we propose, a novel client selection scheme to accelerate the FL convergence by variance reduction. The main idea of is to stratify a compressed model update in order to ensure an excellent grouping effect, and at the same time reduce the cross-client variance by re-allocating the sample chance among different groups based on their diverse heterogeneity. It strikes this convergence acceleration by paying more attention to those client groups with relatively low similarity and then improving the representativeness of the selected subset as much as possible. Theoretically, we demonstrate the critical improvement of the proposed scheme in variance reduction and present equivalence conditions among different client selection methods. We also present the tighter convergence guarantee of the proposed method thanks to the variance reduction. Experimental results confirm the exceeded efficiency of our approach compared to alternatives.</description><identifier>ISSN: 2162-237X</identifier><identifier>ISSN: 2162-2388</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2024.3438843</identifier><identifier>PMID: 39316488</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Client selection ; Convergence ; Costs ; Data models ; federated learning (FL) ; Learning systems ; Monte Carlo methods ; Servers ; stratified sampling ; Training ; variance reduction</subject><ispartof>IEEE transaction on neural networks and learning systems, 2024-09, Vol.PP, p.1-15</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>songdx2000@sjtu.edu.cn ; dehong.gdh@nwpu.edu.cn ; w-zhen@nwpu.edu.cn ; xiaoyanc@nwpu.edu.cn ; libiny@nwpu.edu.cn ; lgshen@sjtu.edu.cn</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10689614$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,27903,27904,54736</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10689614$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/39316488$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Gao, Dehong</creatorcontrib><creatorcontrib>Song, Duanxiao</creatorcontrib><creatorcontrib>Shen, Guangyuan</creatorcontrib><creatorcontrib>Cai, Xiaoyan</creatorcontrib><creatorcontrib>Yang, Libin</creatorcontrib><creatorcontrib>Liu, Gongshen</creatorcontrib><creatorcontrib>Li, Xiaoyong</creatorcontrib><creatorcontrib>Wang, Zhen</creatorcontrib><title>FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-based client selection constitutes a simple yet efficacious approach to the identification of those clients possessing representative gradient information. Despite the extensive body of research on modified selection methodologies, the majority of prior work is predicated upon the assumption of consistently effective clustering. However, raw gradient-based clustering methods are subject to several challenges: 1) poor effectiveness, the raw high-dimensional gradient of a client is too complex to serve as an appropriate feature for grouping, resulting in large intra-cluster distances and 2) fluctuating effectiveness, due to inherent limitations in clustering, the effectiveness can vary significantly, leading to clusters with diverse levels of heterogeneity. In practice, suboptimal and inconsistent clustering effects can result in clusters with low intra-cluster similarity among clients. The selection of clients from such clusters may impede the overall convergence of training. In this article, we propose, a novel client selection scheme to accelerate the FL convergence by variance reduction. The main idea of is to stratify a compressed model update in order to ensure an excellent grouping effect, and at the same time reduce the cross-client variance by re-allocating the sample chance among different groups based on their diverse heterogeneity. It strikes this convergence acceleration by paying more attention to those client groups with relatively low similarity and then improving the representativeness of the selected subset as much as possible. Theoretically, we demonstrate the critical improvement of the proposed scheme in variance reduction and present equivalence conditions among different client selection methods. We also present the tighter convergence guarantee of the proposed method thanks to the variance reduction. Experimental results confirm the exceeded efficiency of our approach compared to alternatives.</description><subject>Client selection</subject><subject>Convergence</subject><subject>Costs</subject><subject>Data models</subject><subject>federated learning (FL)</subject><subject>Learning systems</subject><subject>Monte Carlo methods</subject><subject>Servers</subject><subject>stratified sampling</subject><subject>Training</subject><subject>variance reduction</subject><issn>2162-237X</issn><issn>2162-2388</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpNkF1LwzAUhoMobsz9ARHJpTedaZKmqXdjWBXKvOiE3ZW0PZFoP2bSIfv3Zm4Oz03OIc_7XjwIXYdkFoYkuV8tl1k-o4TyGeNMSs7O0JiGggbUX-enPV6P0NS5D-JHkEjw5BKNWMJCwaUco3UKdb7KH_Ac54NVg9EGarxoDHQDzqGBajB9h1OrWvju7SfWvcWLvnPGDR5pdjhVbsC-BXzaRzNQtjPd-xW60KpxMD2-E_SWPq4Wz0H2-vSymGdBRUk0BKrmQnHOZVwLWpa6FFzVEaVc-58kliWXumKspDXoOibASKlopTSAUhGLGZugu0PvxvZfW3BD0RpXQdOoDvqtK5iXxb2mkHqUHtDK9s5Z0MXGmlbZXRGSYi-1-JVa7KUWR6k-dHvs35Yt1KfIn0IP3BwAAwD_GoVMRMjZDwj2fDQ</recordid><startdate>20240924</startdate><enddate>20240924</enddate><creator>Gao, Dehong</creator><creator>Song, Duanxiao</creator><creator>Shen, Guangyuan</creator><creator>Cai, Xiaoyan</creator><creator>Yang, Libin</creator><creator>Liu, Gongshen</creator><creator>Li, Xiaoyong</creator><creator>Wang, Zhen</creator><general>IEEE</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/songdx2000@sjtu.edu.cn</orcidid><orcidid>https://orcid.org/dehong.gdh@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/w-zhen@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/xiaoyanc@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/libiny@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/lgshen@sjtu.edu.cn</orcidid></search><sort><creationdate>20240924</creationdate><title>FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning</title><author>Gao, Dehong ; Song, Duanxiao ; Shen, Guangyuan ; Cai, Xiaoyan ; Yang, Libin ; Liu, Gongshen ; Li, Xiaoyong ; Wang, Zhen</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c205t-ad46a44487d62bbfb64ad5224fad4978b48fc33b2defd70e30ba2cafeeaa53733</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Client selection</topic><topic>Convergence</topic><topic>Costs</topic><topic>Data models</topic><topic>federated learning (FL)</topic><topic>Learning systems</topic><topic>Monte Carlo methods</topic><topic>Servers</topic><topic>stratified sampling</topic><topic>Training</topic><topic>variance reduction</topic><toplevel>online_resources</toplevel><creatorcontrib>Gao, Dehong</creatorcontrib><creatorcontrib>Song, Duanxiao</creatorcontrib><creatorcontrib>Shen, Guangyuan</creatorcontrib><creatorcontrib>Cai, Xiaoyan</creatorcontrib><creatorcontrib>Yang, Libin</creatorcontrib><creatorcontrib>Liu, Gongshen</creatorcontrib><creatorcontrib>Li, Xiaoyong</creatorcontrib><creatorcontrib>Wang, Zhen</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Gao, Dehong</au><au>Song, Duanxiao</au><au>Shen, Guangyuan</au><au>Cai, Xiaoyan</au><au>Yang, Libin</au><au>Liu, Gongshen</au><au>Li, Xiaoyong</au><au>Wang, Zhen</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2024-09-24</date><risdate>2024</risdate><volume>PP</volume><spage>1</spage><epage>15</epage><pages>1-15</pages><issn>2162-237X</issn><issn>2162-2388</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-based client selection constitutes a simple yet efficacious approach to the identification of those clients possessing representative gradient information. Despite the extensive body of research on modified selection methodologies, the majority of prior work is predicated upon the assumption of consistently effective clustering. However, raw gradient-based clustering methods are subject to several challenges: 1) poor effectiveness, the raw high-dimensional gradient of a client is too complex to serve as an appropriate feature for grouping, resulting in large intra-cluster distances and 2) fluctuating effectiveness, due to inherent limitations in clustering, the effectiveness can vary significantly, leading to clusters with diverse levels of heterogeneity. In practice, suboptimal and inconsistent clustering effects can result in clusters with low intra-cluster similarity among clients. The selection of clients from such clusters may impede the overall convergence of training. In this article, we propose, a novel client selection scheme to accelerate the FL convergence by variance reduction. The main idea of is to stratify a compressed model update in order to ensure an excellent grouping effect, and at the same time reduce the cross-client variance by re-allocating the sample chance among different groups based on their diverse heterogeneity. It strikes this convergence acceleration by paying more attention to those client groups with relatively low similarity and then improving the representativeness of the selected subset as much as possible. Theoretically, we demonstrate the critical improvement of the proposed scheme in variance reduction and present equivalence conditions among different client selection methods. We also present the tighter convergence guarantee of the proposed method thanks to the variance reduction. Experimental results confirm the exceeded efficiency of our approach compared to alternatives.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>39316488</pmid><doi>10.1109/TNNLS.2024.3438843</doi><tpages>15</tpages><orcidid>https://orcid.org/songdx2000@sjtu.edu.cn</orcidid><orcidid>https://orcid.org/dehong.gdh@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/w-zhen@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/xiaoyanc@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/libiny@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/lgshen@sjtu.edu.cn</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier ISSN: 2162-237X
ispartof IEEE transaction on neural networks and learning systems, 2024-09, Vol.PP, p.1-15
issn 2162-237X
2162-2388
2162-2388
language eng
recordid cdi_proquest_miscellaneous_3109420212
source IEEE Electronic Library (IEL)
subjects Client selection
Convergence
Costs
Data models
federated learning (FL)
Learning systems
Monte Carlo methods
Servers
stratified sampling
Training
variance reduction
title FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-26T03%3A12%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=FedSTS:%20A%20Stratified%20Client%20Selection%20Framework%20for%20Consistently%20Fast%20Federated%20Learning&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Gao,%20Dehong&rft.date=2024-09-24&rft.volume=PP&rft.spage=1&rft.epage=15&rft.pages=1-15&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2024.3438843&rft_dat=%3Cproquest_RIE%3E3109420212%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3109420212&rft_id=info:pmid/39316488&rft_ieee_id=10689614&rfr_iscdi=true