FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning
In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-bas...
Gespeichert in:
Veröffentlicht in: | IEEE transaction on neural networks and learning systems 2024-09, Vol.PP, p.1-15 |
---|---|
Hauptverfasser: | , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 15 |
---|---|
container_issue | |
container_start_page | 1 |
container_title | IEEE transaction on neural networks and learning systems |
container_volume | PP |
creator | Gao, Dehong Song, Duanxiao Shen, Guangyuan Cai, Xiaoyan Yang, Libin Liu, Gongshen Li, Xiaoyong Wang, Zhen |
description | In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-based client selection constitutes a simple yet efficacious approach to the identification of those clients possessing representative gradient information. Despite the extensive body of research on modified selection methodologies, the majority of prior work is predicated upon the assumption of consistently effective clustering. However, raw gradient-based clustering methods are subject to several challenges: 1) poor effectiveness, the raw high-dimensional gradient of a client is too complex to serve as an appropriate feature for grouping, resulting in large intra-cluster distances and 2) fluctuating effectiveness, due to inherent limitations in clustering, the effectiveness can vary significantly, leading to clusters with diverse levels of heterogeneity. In practice, suboptimal and inconsistent clustering effects can result in clusters with low intra-cluster similarity among clients. The selection of clients from such clusters may impede the overall convergence of training. In this article, we propose, a novel client selection scheme to accelerate the FL convergence by variance reduction. The main idea of is to stratify a compressed model update in order to ensure an excellent grouping effect, and at the same time reduce the cross-client variance by re-allocating the sample chance among different groups based on their diverse heterogeneity. It strikes this convergence acceleration by paying more attention to those client groups with relatively low similarity and then improving the representativeness of the selected subset as much as possible. Theoretically, we demonstrate the critical improvement of the proposed scheme in variance reduction and present equivalence conditions among different client selection methods. We also present the tighter convergence guarantee of the proposed method thanks to the variance reduction. Experimental results confirm the exceeded efficiency of our approach compared to alternatives. |
doi_str_mv | 10.1109/TNNLS.2024.3438843 |
format | Article |
fullrecord | <record><control><sourceid>proquest_RIE</sourceid><recordid>TN_cdi_proquest_miscellaneous_3109420212</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10689614</ieee_id><sourcerecordid>3109420212</sourcerecordid><originalsourceid>FETCH-LOGICAL-c205t-ad46a44487d62bbfb64ad5224fad4978b48fc33b2defd70e30ba2cafeeaa53733</originalsourceid><addsrcrecordid>eNpNkF1LwzAUhoMobsz9ARHJpTedaZKmqXdjWBXKvOiE3ZW0PZFoP2bSIfv3Zm4Oz03OIc_7XjwIXYdkFoYkuV8tl1k-o4TyGeNMSs7O0JiGggbUX-enPV6P0NS5D-JHkEjw5BKNWMJCwaUco3UKdb7KH_Ac54NVg9EGarxoDHQDzqGBajB9h1OrWvju7SfWvcWLvnPGDR5pdjhVbsC-BXzaRzNQtjPd-xW60KpxMD2-E_SWPq4Wz0H2-vSymGdBRUk0BKrmQnHOZVwLWpa6FFzVEaVc-58kliWXumKspDXoOibASKlopTSAUhGLGZugu0PvxvZfW3BD0RpXQdOoDvqtK5iXxb2mkHqUHtDK9s5Z0MXGmlbZXRGSYi-1-JVa7KUWR6k-dHvs35Yt1KfIn0IP3BwAAwD_GoVMRMjZDwj2fDQ</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>3109420212</pqid></control><display><type>article</type><title>FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning</title><source>IEEE Electronic Library (IEL)</source><creator>Gao, Dehong ; Song, Duanxiao ; Shen, Guangyuan ; Cai, Xiaoyan ; Yang, Libin ; Liu, Gongshen ; Li, Xiaoyong ; Wang, Zhen</creator><creatorcontrib>Gao, Dehong ; Song, Duanxiao ; Shen, Guangyuan ; Cai, Xiaoyan ; Yang, Libin ; Liu, Gongshen ; Li, Xiaoyong ; Wang, Zhen</creatorcontrib><description>In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-based client selection constitutes a simple yet efficacious approach to the identification of those clients possessing representative gradient information. Despite the extensive body of research on modified selection methodologies, the majority of prior work is predicated upon the assumption of consistently effective clustering. However, raw gradient-based clustering methods are subject to several challenges: 1) poor effectiveness, the raw high-dimensional gradient of a client is too complex to serve as an appropriate feature for grouping, resulting in large intra-cluster distances and 2) fluctuating effectiveness, due to inherent limitations in clustering, the effectiveness can vary significantly, leading to clusters with diverse levels of heterogeneity. In practice, suboptimal and inconsistent clustering effects can result in clusters with low intra-cluster similarity among clients. The selection of clients from such clusters may impede the overall convergence of training. In this article, we propose, a novel client selection scheme to accelerate the FL convergence by variance reduction. The main idea of is to stratify a compressed model update in order to ensure an excellent grouping effect, and at the same time reduce the cross-client variance by re-allocating the sample chance among different groups based on their diverse heterogeneity. It strikes this convergence acceleration by paying more attention to those client groups with relatively low similarity and then improving the representativeness of the selected subset as much as possible. Theoretically, we demonstrate the critical improvement of the proposed scheme in variance reduction and present equivalence conditions among different client selection methods. We also present the tighter convergence guarantee of the proposed method thanks to the variance reduction. Experimental results confirm the exceeded efficiency of our approach compared to alternatives.</description><identifier>ISSN: 2162-237X</identifier><identifier>ISSN: 2162-2388</identifier><identifier>EISSN: 2162-2388</identifier><identifier>DOI: 10.1109/TNNLS.2024.3438843</identifier><identifier>PMID: 39316488</identifier><identifier>CODEN: ITNNAL</identifier><language>eng</language><publisher>United States: IEEE</publisher><subject>Client selection ; Convergence ; Costs ; Data models ; federated learning (FL) ; Learning systems ; Monte Carlo methods ; Servers ; stratified sampling ; Training ; variance reduction</subject><ispartof>IEEE transaction on neural networks and learning systems, 2024-09, Vol.PP, p.1-15</ispartof><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>songdx2000@sjtu.edu.cn ; dehong.gdh@nwpu.edu.cn ; w-zhen@nwpu.edu.cn ; xiaoyanc@nwpu.edu.cn ; libiny@nwpu.edu.cn ; lgshen@sjtu.edu.cn</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10689614$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,776,780,792,27903,27904,54736</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10689614$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc><backlink>$$Uhttps://www.ncbi.nlm.nih.gov/pubmed/39316488$$D View this record in MEDLINE/PubMed$$Hfree_for_read</backlink></links><search><creatorcontrib>Gao, Dehong</creatorcontrib><creatorcontrib>Song, Duanxiao</creatorcontrib><creatorcontrib>Shen, Guangyuan</creatorcontrib><creatorcontrib>Cai, Xiaoyan</creatorcontrib><creatorcontrib>Yang, Libin</creatorcontrib><creatorcontrib>Liu, Gongshen</creatorcontrib><creatorcontrib>Li, Xiaoyong</creatorcontrib><creatorcontrib>Wang, Zhen</creatorcontrib><title>FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning</title><title>IEEE transaction on neural networks and learning systems</title><addtitle>TNNLS</addtitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><description>In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-based client selection constitutes a simple yet efficacious approach to the identification of those clients possessing representative gradient information. Despite the extensive body of research on modified selection methodologies, the majority of prior work is predicated upon the assumption of consistently effective clustering. However, raw gradient-based clustering methods are subject to several challenges: 1) poor effectiveness, the raw high-dimensional gradient of a client is too complex to serve as an appropriate feature for grouping, resulting in large intra-cluster distances and 2) fluctuating effectiveness, due to inherent limitations in clustering, the effectiveness can vary significantly, leading to clusters with diverse levels of heterogeneity. In practice, suboptimal and inconsistent clustering effects can result in clusters with low intra-cluster similarity among clients. The selection of clients from such clusters may impede the overall convergence of training. In this article, we propose, a novel client selection scheme to accelerate the FL convergence by variance reduction. The main idea of is to stratify a compressed model update in order to ensure an excellent grouping effect, and at the same time reduce the cross-client variance by re-allocating the sample chance among different groups based on their diverse heterogeneity. It strikes this convergence acceleration by paying more attention to those client groups with relatively low similarity and then improving the representativeness of the selected subset as much as possible. Theoretically, we demonstrate the critical improvement of the proposed scheme in variance reduction and present equivalence conditions among different client selection methods. We also present the tighter convergence guarantee of the proposed method thanks to the variance reduction. Experimental results confirm the exceeded efficiency of our approach compared to alternatives.</description><subject>Client selection</subject><subject>Convergence</subject><subject>Costs</subject><subject>Data models</subject><subject>federated learning (FL)</subject><subject>Learning systems</subject><subject>Monte Carlo methods</subject><subject>Servers</subject><subject>stratified sampling</subject><subject>Training</subject><subject>variance reduction</subject><issn>2162-237X</issn><issn>2162-2388</issn><issn>2162-2388</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNpNkF1LwzAUhoMobsz9ARHJpTedaZKmqXdjWBXKvOiE3ZW0PZFoP2bSIfv3Zm4Oz03OIc_7XjwIXYdkFoYkuV8tl1k-o4TyGeNMSs7O0JiGggbUX-enPV6P0NS5D-JHkEjw5BKNWMJCwaUco3UKdb7KH_Ac54NVg9EGarxoDHQDzqGBajB9h1OrWvju7SfWvcWLvnPGDR5pdjhVbsC-BXzaRzNQtjPd-xW60KpxMD2-E_SWPq4Wz0H2-vSymGdBRUk0BKrmQnHOZVwLWpa6FFzVEaVc-58kliWXumKspDXoOibASKlopTSAUhGLGZugu0PvxvZfW3BD0RpXQdOoDvqtK5iXxb2mkHqUHtDK9s5Z0MXGmlbZXRGSYi-1-JVa7KUWR6k-dHvs35Yt1KfIn0IP3BwAAwD_GoVMRMjZDwj2fDQ</recordid><startdate>20240924</startdate><enddate>20240924</enddate><creator>Gao, Dehong</creator><creator>Song, Duanxiao</creator><creator>Shen, Guangyuan</creator><creator>Cai, Xiaoyan</creator><creator>Yang, Libin</creator><creator>Liu, Gongshen</creator><creator>Li, Xiaoyong</creator><creator>Wang, Zhen</creator><general>IEEE</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><scope>NPM</scope><scope>AAYXX</scope><scope>CITATION</scope><scope>7X8</scope><orcidid>https://orcid.org/songdx2000@sjtu.edu.cn</orcidid><orcidid>https://orcid.org/dehong.gdh@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/w-zhen@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/xiaoyanc@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/libiny@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/lgshen@sjtu.edu.cn</orcidid></search><sort><creationdate>20240924</creationdate><title>FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning</title><author>Gao, Dehong ; Song, Duanxiao ; Shen, Guangyuan ; Cai, Xiaoyan ; Yang, Libin ; Liu, Gongshen ; Li, Xiaoyong ; Wang, Zhen</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c205t-ad46a44487d62bbfb64ad5224fad4978b48fc33b2defd70e30ba2cafeeaa53733</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Client selection</topic><topic>Convergence</topic><topic>Costs</topic><topic>Data models</topic><topic>federated learning (FL)</topic><topic>Learning systems</topic><topic>Monte Carlo methods</topic><topic>Servers</topic><topic>stratified sampling</topic><topic>Training</topic><topic>variance reduction</topic><toplevel>online_resources</toplevel><creatorcontrib>Gao, Dehong</creatorcontrib><creatorcontrib>Song, Duanxiao</creatorcontrib><creatorcontrib>Shen, Guangyuan</creatorcontrib><creatorcontrib>Cai, Xiaoyan</creatorcontrib><creatorcontrib>Yang, Libin</creatorcontrib><creatorcontrib>Liu, Gongshen</creatorcontrib><creatorcontrib>Li, Xiaoyong</creatorcontrib><creatorcontrib>Wang, Zhen</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><collection>PubMed</collection><collection>CrossRef</collection><collection>MEDLINE - Academic</collection><jtitle>IEEE transaction on neural networks and learning systems</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Gao, Dehong</au><au>Song, Duanxiao</au><au>Shen, Guangyuan</au><au>Cai, Xiaoyan</au><au>Yang, Libin</au><au>Liu, Gongshen</au><au>Li, Xiaoyong</au><au>Wang, Zhen</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning</atitle><jtitle>IEEE transaction on neural networks and learning systems</jtitle><stitle>TNNLS</stitle><addtitle>IEEE Trans Neural Netw Learn Syst</addtitle><date>2024-09-24</date><risdate>2024</risdate><volume>PP</volume><spage>1</spage><epage>15</epage><pages>1-15</pages><issn>2162-237X</issn><issn>2162-2388</issn><eissn>2162-2388</eissn><coden>ITNNAL</coden><abstract>In this article, we investigate random client selection in the context of horizontal federated learning (FL), whereby only a randomly selected subset of clients transmit their model updates to the server instead of yielding all clients involved. Many researchers have demonstrated that clustering-based client selection constitutes a simple yet efficacious approach to the identification of those clients possessing representative gradient information. Despite the extensive body of research on modified selection methodologies, the majority of prior work is predicated upon the assumption of consistently effective clustering. However, raw gradient-based clustering methods are subject to several challenges: 1) poor effectiveness, the raw high-dimensional gradient of a client is too complex to serve as an appropriate feature for grouping, resulting in large intra-cluster distances and 2) fluctuating effectiveness, due to inherent limitations in clustering, the effectiveness can vary significantly, leading to clusters with diverse levels of heterogeneity. In practice, suboptimal and inconsistent clustering effects can result in clusters with low intra-cluster similarity among clients. The selection of clients from such clusters may impede the overall convergence of training. In this article, we propose, a novel client selection scheme to accelerate the FL convergence by variance reduction. The main idea of is to stratify a compressed model update in order to ensure an excellent grouping effect, and at the same time reduce the cross-client variance by re-allocating the sample chance among different groups based on their diverse heterogeneity. It strikes this convergence acceleration by paying more attention to those client groups with relatively low similarity and then improving the representativeness of the selected subset as much as possible. Theoretically, we demonstrate the critical improvement of the proposed scheme in variance reduction and present equivalence conditions among different client selection methods. We also present the tighter convergence guarantee of the proposed method thanks to the variance reduction. Experimental results confirm the exceeded efficiency of our approach compared to alternatives.</abstract><cop>United States</cop><pub>IEEE</pub><pmid>39316488</pmid><doi>10.1109/TNNLS.2024.3438843</doi><tpages>15</tpages><orcidid>https://orcid.org/songdx2000@sjtu.edu.cn</orcidid><orcidid>https://orcid.org/dehong.gdh@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/w-zhen@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/xiaoyanc@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/libiny@nwpu.edu.cn</orcidid><orcidid>https://orcid.org/lgshen@sjtu.edu.cn</orcidid></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | ISSN: 2162-237X |
ispartof | IEEE transaction on neural networks and learning systems, 2024-09, Vol.PP, p.1-15 |
issn | 2162-237X 2162-2388 2162-2388 |
language | eng |
recordid | cdi_proquest_miscellaneous_3109420212 |
source | IEEE Electronic Library (IEL) |
subjects | Client selection Convergence Costs Data models federated learning (FL) Learning systems Monte Carlo methods Servers stratified sampling Training variance reduction |
title | FedSTS: A Stratified Client Selection Framework for Consistently Fast Federated Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-26T03%3A12%3A04IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=FedSTS:%20A%20Stratified%20Client%20Selection%20Framework%20for%20Consistently%20Fast%20Federated%20Learning&rft.jtitle=IEEE%20transaction%20on%20neural%20networks%20and%20learning%20systems&rft.au=Gao,%20Dehong&rft.date=2024-09-24&rft.volume=PP&rft.spage=1&rft.epage=15&rft.pages=1-15&rft.issn=2162-237X&rft.eissn=2162-2388&rft.coden=ITNNAL&rft_id=info:doi/10.1109/TNNLS.2024.3438843&rft_dat=%3Cproquest_RIE%3E3109420212%3C/proquest_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=3109420212&rft_id=info:pmid/39316488&rft_ieee_id=10689614&rfr_iscdi=true |