FedDCT: Federated Learning of Large Convolutional Neural Networks on Resource Constrained Devices using Divide and Collaborative Training
We introduce FedDCT, a novel distributed learning paradigm that enables the usage of large, high-performance CNNs on resource-limited edge devices. As opposed to traditional FL approaches, which require each client to train the full-size neural network independently during each training round, the p...
Gespeichert in:
Veröffentlicht in: | arXiv.org 2023-09 |
---|---|
Hauptverfasser: | , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | arXiv.org |
container_volume | |
creator | Nguyen, Quan Pham, Hieu H Kok-Seng Wong Phi Le Nguyen Truong Thao Nguyen Do, Minh N |
description | We introduce FedDCT, a novel distributed learning paradigm that enables the usage of large, high-performance CNNs on resource-limited edge devices. As opposed to traditional FL approaches, which require each client to train the full-size neural network independently during each training round, the proposed FedDCT allows a cluster of several clients to collaboratively train a large deep learning model by dividing it into an ensemble of several small sub-models and train them on multiple devices in parallel while maintaining privacy. In this collaborative training process, clients from the same cluster can also learn from each other, further improving their ensemble performance. In the aggregation stage, the server takes a weighted average of all the ensemble models trained by all the clusters. FedDCT reduces the memory requirements and allows low-end devices to participate in FL. We empirically conduct extensive experiments on standardized datasets, including CIFAR-10, CIFAR-100, and two real-world medical datasets HAM10000 and VAIPE. Experimental results show that FedDCT outperforms a set of current SOTA FL methods with interesting convergence behaviors. Furthermore, compared to other existing approaches, FedDCT achieves higher accuracy and substantially reduces the number of communication rounds (with \(4-8\) times fewer memory requirements) to achieve the desired accuracy on the testing dataset without incurring any extra training cost on the server side. |
doi_str_mv | 10.48550/arxiv.2211.10948 |
format | Article |
fullrecord | <record><control><sourceid>proquest_arxiv</sourceid><recordid>TN_cdi_arxiv_primary_2211_10948</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2738700467</sourcerecordid><originalsourceid>FETCH-LOGICAL-a958-8b607173a1cbc5081ced4034b8969f15db8a5a1506d114cc373936998d240af93</originalsourceid><addsrcrecordid>eNotkN1KwzAAhYMgOOYewCsDXnfmt028k86pUBSk9yVt0pFZm5m0VR_BtzbrvDo3H-dwPgCuMFozwTm6Vf7bTmtCMF5jJJk4AwtCKU4EI-QCrELYI4RImhHO6QL8bo3e5OUdjGm8GoyGhVG-t_0OuhYWyu8MzF0_uW4crOtVB1_M6OcYvpx_D9D18M0EN_pmJsPgle1jz8ZMtjEBjuFYtrGT1QaqXkeo61Tt4pqdDCyPeCQuwXmrumBW_7kE5fahzJ-S4vXxOb8vEiW5SESdogxnVOGmbjgSuDGaIcpqIVPZYq5robjCHKUaY9Y0NKOSplIKTRhSraRLcH2qnT1VB28_lP-pjr6q2Vckbk7EwbvP0YSh2sdz8XmoSEZFhhBLM_oHnipukQ</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2738700467</pqid></control><display><type>article</type><title>FedDCT: Federated Learning of Large Convolutional Neural Networks on Resource Constrained Devices using Divide and Collaborative Training</title><source>arXiv.org</source><source>Free E- Journals</source><creator>Nguyen, Quan ; Pham, Hieu H ; Kok-Seng Wong ; Phi Le Nguyen ; Truong Thao Nguyen ; Do, Minh N</creator><creatorcontrib>Nguyen, Quan ; Pham, Hieu H ; Kok-Seng Wong ; Phi Le Nguyen ; Truong Thao Nguyen ; Do, Minh N</creatorcontrib><description>We introduce FedDCT, a novel distributed learning paradigm that enables the usage of large, high-performance CNNs on resource-limited edge devices. As opposed to traditional FL approaches, which require each client to train the full-size neural network independently during each training round, the proposed FedDCT allows a cluster of several clients to collaboratively train a large deep learning model by dividing it into an ensemble of several small sub-models and train them on multiple devices in parallel while maintaining privacy. In this collaborative training process, clients from the same cluster can also learn from each other, further improving their ensemble performance. In the aggregation stage, the server takes a weighted average of all the ensemble models trained by all the clusters. FedDCT reduces the memory requirements and allows low-end devices to participate in FL. We empirically conduct extensive experiments on standardized datasets, including CIFAR-10, CIFAR-100, and two real-world medical datasets HAM10000 and VAIPE. Experimental results show that FedDCT outperforms a set of current SOTA FL methods with interesting convergence behaviors. Furthermore, compared to other existing approaches, FedDCT achieves higher accuracy and substantially reduces the number of communication rounds (with \(4-8\) times fewer memory requirements) to achieve the desired accuracy on the testing dataset without incurring any extra training cost on the server side.</description><identifier>EISSN: 2331-8422</identifier><identifier>DOI: 10.48550/arxiv.2211.10948</identifier><language>eng</language><publisher>Ithaca: Cornell University Library, arXiv.org</publisher><subject>Accuracy ; Artificial neural networks ; Clients ; Clusters ; Computer Science - Computer Vision and Pattern Recognition ; Datasets ; Deep learning ; Federated learning ; Neural networks ; Servers</subject><ispartof>arXiv.org, 2023-09</ispartof><rights>2023. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.</rights><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,781,785,886,27930</link.rule.ids><backlink>$$Uhttps://doi.org/10.48550/arXiv.2211.10948$$DView paper in arXiv$$Hfree_for_read</backlink><backlink>$$Uhttps://doi.org/10.1109/TNSM.2023.3314066$$DView published paper (Access to full text may be restricted)$$Hfree_for_read</backlink></links><search><creatorcontrib>Nguyen, Quan</creatorcontrib><creatorcontrib>Pham, Hieu H</creatorcontrib><creatorcontrib>Kok-Seng Wong</creatorcontrib><creatorcontrib>Phi Le Nguyen</creatorcontrib><creatorcontrib>Truong Thao Nguyen</creatorcontrib><creatorcontrib>Do, Minh N</creatorcontrib><title>FedDCT: Federated Learning of Large Convolutional Neural Networks on Resource Constrained Devices using Divide and Collaborative Training</title><title>arXiv.org</title><description>We introduce FedDCT, a novel distributed learning paradigm that enables the usage of large, high-performance CNNs on resource-limited edge devices. As opposed to traditional FL approaches, which require each client to train the full-size neural network independently during each training round, the proposed FedDCT allows a cluster of several clients to collaboratively train a large deep learning model by dividing it into an ensemble of several small sub-models and train them on multiple devices in parallel while maintaining privacy. In this collaborative training process, clients from the same cluster can also learn from each other, further improving their ensemble performance. In the aggregation stage, the server takes a weighted average of all the ensemble models trained by all the clusters. FedDCT reduces the memory requirements and allows low-end devices to participate in FL. We empirically conduct extensive experiments on standardized datasets, including CIFAR-10, CIFAR-100, and two real-world medical datasets HAM10000 and VAIPE. Experimental results show that FedDCT outperforms a set of current SOTA FL methods with interesting convergence behaviors. Furthermore, compared to other existing approaches, FedDCT achieves higher accuracy and substantially reduces the number of communication rounds (with \(4-8\) times fewer memory requirements) to achieve the desired accuracy on the testing dataset without incurring any extra training cost on the server side.</description><subject>Accuracy</subject><subject>Artificial neural networks</subject><subject>Clients</subject><subject>Clusters</subject><subject>Computer Science - Computer Vision and Pattern Recognition</subject><subject>Datasets</subject><subject>Deep learning</subject><subject>Federated learning</subject><subject>Neural networks</subject><subject>Servers</subject><issn>2331-8422</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><sourceid>ABUWG</sourceid><sourceid>AFKRA</sourceid><sourceid>AZQEC</sourceid><sourceid>BENPR</sourceid><sourceid>CCPQU</sourceid><sourceid>DWQXO</sourceid><sourceid>GOX</sourceid><recordid>eNotkN1KwzAAhYMgOOYewCsDXnfmt028k86pUBSk9yVt0pFZm5m0VR_BtzbrvDo3H-dwPgCuMFozwTm6Vf7bTmtCMF5jJJk4AwtCKU4EI-QCrELYI4RImhHO6QL8bo3e5OUdjGm8GoyGhVG-t_0OuhYWyu8MzF0_uW4crOtVB1_M6OcYvpx_D9D18M0EN_pmJsPgle1jz8ZMtjEBjuFYtrGT1QaqXkeo61Tt4pqdDCyPeCQuwXmrumBW_7kE5fahzJ-S4vXxOb8vEiW5SESdogxnVOGmbjgSuDGaIcpqIVPZYq5robjCHKUaY9Y0NKOSplIKTRhSraRLcH2qnT1VB28_lP-pjr6q2Vckbk7EwbvP0YSh2sdz8XmoSEZFhhBLM_oHnipukQ</recordid><startdate>20230918</startdate><enddate>20230918</enddate><creator>Nguyen, Quan</creator><creator>Pham, Hieu H</creator><creator>Kok-Seng Wong</creator><creator>Phi Le Nguyen</creator><creator>Truong Thao Nguyen</creator><creator>Do, Minh N</creator><general>Cornell University Library, arXiv.org</general><scope>8FE</scope><scope>8FG</scope><scope>ABJCF</scope><scope>ABUWG</scope><scope>AFKRA</scope><scope>AZQEC</scope><scope>BENPR</scope><scope>BGLVJ</scope><scope>CCPQU</scope><scope>DWQXO</scope><scope>HCIFZ</scope><scope>L6V</scope><scope>M7S</scope><scope>PIMPY</scope><scope>PQEST</scope><scope>PQQKQ</scope><scope>PQUKI</scope><scope>PRINS</scope><scope>PTHSS</scope><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20230918</creationdate><title>FedDCT: Federated Learning of Large Convolutional Neural Networks on Resource Constrained Devices using Divide and Collaborative Training</title><author>Nguyen, Quan ; Pham, Hieu H ; Kok-Seng Wong ; Phi Le Nguyen ; Truong Thao Nguyen ; Do, Minh N</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a958-8b607173a1cbc5081ced4034b8969f15db8a5a1506d114cc373936998d240af93</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Accuracy</topic><topic>Artificial neural networks</topic><topic>Clients</topic><topic>Clusters</topic><topic>Computer Science - Computer Vision and Pattern Recognition</topic><topic>Datasets</topic><topic>Deep learning</topic><topic>Federated learning</topic><topic>Neural networks</topic><topic>Servers</topic><toplevel>online_resources</toplevel><creatorcontrib>Nguyen, Quan</creatorcontrib><creatorcontrib>Pham, Hieu H</creatorcontrib><creatorcontrib>Kok-Seng Wong</creatorcontrib><creatorcontrib>Phi Le Nguyen</creatorcontrib><creatorcontrib>Truong Thao Nguyen</creatorcontrib><creatorcontrib>Do, Minh N</creatorcontrib><collection>ProQuest SciTech Collection</collection><collection>ProQuest Technology Collection</collection><collection>Materials Science & Engineering Collection</collection><collection>ProQuest Central (Alumni Edition)</collection><collection>ProQuest Central UK/Ireland</collection><collection>ProQuest Central Essentials</collection><collection>ProQuest Central</collection><collection>Technology Collection</collection><collection>ProQuest One Community College</collection><collection>ProQuest Central Korea</collection><collection>SciTech Premium Collection</collection><collection>ProQuest Engineering Collection</collection><collection>Engineering Database</collection><collection>Publicly Available Content Database</collection><collection>ProQuest One Academic Eastern Edition (DO NOT USE)</collection><collection>ProQuest One Academic</collection><collection>ProQuest One Academic UKI Edition</collection><collection>ProQuest Central China</collection><collection>Engineering Collection</collection><collection>arXiv Computer Science</collection><collection>arXiv.org</collection><jtitle>arXiv.org</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Nguyen, Quan</au><au>Pham, Hieu H</au><au>Kok-Seng Wong</au><au>Phi Le Nguyen</au><au>Truong Thao Nguyen</au><au>Do, Minh N</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>FedDCT: Federated Learning of Large Convolutional Neural Networks on Resource Constrained Devices using Divide and Collaborative Training</atitle><jtitle>arXiv.org</jtitle><date>2023-09-18</date><risdate>2023</risdate><eissn>2331-8422</eissn><abstract>We introduce FedDCT, a novel distributed learning paradigm that enables the usage of large, high-performance CNNs on resource-limited edge devices. As opposed to traditional FL approaches, which require each client to train the full-size neural network independently during each training round, the proposed FedDCT allows a cluster of several clients to collaboratively train a large deep learning model by dividing it into an ensemble of several small sub-models and train them on multiple devices in parallel while maintaining privacy. In this collaborative training process, clients from the same cluster can also learn from each other, further improving their ensemble performance. In the aggregation stage, the server takes a weighted average of all the ensemble models trained by all the clusters. FedDCT reduces the memory requirements and allows low-end devices to participate in FL. We empirically conduct extensive experiments on standardized datasets, including CIFAR-10, CIFAR-100, and two real-world medical datasets HAM10000 and VAIPE. Experimental results show that FedDCT outperforms a set of current SOTA FL methods with interesting convergence behaviors. Furthermore, compared to other existing approaches, FedDCT achieves higher accuracy and substantially reduces the number of communication rounds (with \(4-8\) times fewer memory requirements) to achieve the desired accuracy on the testing dataset without incurring any extra training cost on the server side.</abstract><cop>Ithaca</cop><pub>Cornell University Library, arXiv.org</pub><doi>10.48550/arxiv.2211.10948</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext |
identifier | EISSN: 2331-8422 |
ispartof | arXiv.org, 2023-09 |
issn | 2331-8422 |
language | eng |
recordid | cdi_arxiv_primary_2211_10948 |
source | arXiv.org; Free E- Journals |
subjects | Accuracy Artificial neural networks Clients Clusters Computer Science - Computer Vision and Pattern Recognition Datasets Deep learning Federated learning Neural networks Servers |
title | FedDCT: Federated Learning of Large Convolutional Neural Networks on Resource Constrained Devices using Divide and Collaborative Training |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-14T18%3A17%3A35IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_arxiv&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=FedDCT:%20Federated%20Learning%20of%20Large%20Convolutional%20Neural%20Networks%20on%20Resource%20Constrained%20Devices%20using%20Divide%20and%20Collaborative%20Training&rft.jtitle=arXiv.org&rft.au=Nguyen,%20Quan&rft.date=2023-09-18&rft.eissn=2331-8422&rft_id=info:doi/10.48550/arxiv.2211.10948&rft_dat=%3Cproquest_arxiv%3E2738700467%3C/proquest_arxiv%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2738700467&rft_id=info:pmid/&rfr_iscdi=true |