Multi-Job Intelligent Scheduling with Cross-Device Federated Learning

Recent years have witnessed a large amount of decentralized data in various (edge) devices of end-users, while the decentralized data aggregation remains complicated for machine learning jobs because of regulations and laws. As a practical approach to handling decentralized data, Federated Learning...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Hauptverfasser: Liu, Ji, Jia, Juncheng, Ma, Beichen, Zhou, Chendi, Zhou, Jingbo, Zhou, Yang, Dai, Huaiyu, Dou, Dejing
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page
container_issue
container_start_page
container_title
container_volume
creator Liu, Ji
Jia, Juncheng
Ma, Beichen
Zhou, Chendi
Zhou, Jingbo
Zhou, Yang
Dai, Huaiyu
Dou, Dejing
description Recent years have witnessed a large amount of decentralized data in various (edge) devices of end-users, while the decentralized data aggregation remains complicated for machine learning jobs because of regulations and laws. As a practical approach to handling decentralized data, Federated Learning (FL) enables collaborative global machine learning model training without sharing sensitive raw data. The servers schedule devices to jobs within the training process of FL. In contrast, device scheduling with multiple jobs in FL remains a critical and open problem. In this paper, we propose a novel multi-job FL framework, which enables the training process of multiple jobs in parallel. The multi-job FL framework is composed of a system model and a scheduling method. The system model enables a parallel training process of multiple jobs, with a cost model based on the data fairness and the training time of diverse devices during the parallel training process. We propose a novel intelligent scheduling approach based on multiple scheduling methods, including an original reinforcement learning-based scheduling method and an original Bayesian optimization-based scheduling method, which corresponds to a small cost while scheduling devices to multiple jobs. We conduct extensive experimentation with diverse jobs and datasets. The experimental results reveal that our proposed approaches significantly outperform baseline approaches in terms of training time (up to 12.73 times faster) and accuracy (up to 46.4% higher).
doi_str_mv 10.48550/arxiv.2211.13430
format Article
fullrecord <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2211_13430</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2211_13430</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-97a24054f873c7d86900eb97c8acc64202e32a6cdbaac3bf475e166f32b067cd3</originalsourceid><addsrcrecordid>eNotz71OwzAYhWEvDKhwAUz4Bhz8FzsZUWihKIiB7tFn-0tryaTIcQvcPVA6neXVkR5CbgSvdFPX_A7yVzxWUgpRCaUVvyTLl0MqkT3vHV1PBVOKW5wKffM7DIcUpy39jGVHu7yfZ_aAx-iRrjBghoKB9gh5-o2uyMUIacbr8y7IZrXcdE-sf31cd_c9A2M5ay1IzWs9NlZ5GxrTco6utb4B742WXKKSYHxwAF65UdsahTGjko4b64NakNv_25Nj-MjxHfL38OcZTh71A3PURWg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Multi-Job Intelligent Scheduling with Cross-Device Federated Learning</title><source>arXiv.org</source><creator>Liu, Ji ; Jia, Juncheng ; Ma, Beichen ; Zhou, Chendi ; Zhou, Jingbo ; Zhou, Yang ; Dai, Huaiyu ; Dou, Dejing</creator><creatorcontrib>Liu, Ji ; Jia, Juncheng ; Ma, Beichen ; Zhou, Chendi ; Zhou, Jingbo ; Zhou, Yang ; Dai, Huaiyu ; Dou, Dejing</creatorcontrib><description>Recent years have witnessed a large amount of decentralized data in various (edge) devices of end-users, while the decentralized data aggregation remains complicated for machine learning jobs because of regulations and laws. As a practical approach to handling decentralized data, Federated Learning (FL) enables collaborative global machine learning model training without sharing sensitive raw data. The servers schedule devices to jobs within the training process of FL. In contrast, device scheduling with multiple jobs in FL remains a critical and open problem. In this paper, we propose a novel multi-job FL framework, which enables the training process of multiple jobs in parallel. The multi-job FL framework is composed of a system model and a scheduling method. The system model enables a parallel training process of multiple jobs, with a cost model based on the data fairness and the training time of diverse devices during the parallel training process. We propose a novel intelligent scheduling approach based on multiple scheduling methods, including an original reinforcement learning-based scheduling method and an original Bayesian optimization-based scheduling method, which corresponds to a small cost while scheduling devices to multiple jobs. We conduct extensive experimentation with diverse jobs and datasets. The experimental results reveal that our proposed approaches significantly outperform baseline approaches in terms of training time (up to 12.73 times faster) and accuracy (up to 46.4% higher).</description><identifier>DOI: 10.48550/arxiv.2211.13430</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Distributed, Parallel, and Cluster Computing ; Computer Science - Learning</subject><creationdate>2022-11</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2211.13430$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2211.13430$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Liu, Ji</creatorcontrib><creatorcontrib>Jia, Juncheng</creatorcontrib><creatorcontrib>Ma, Beichen</creatorcontrib><creatorcontrib>Zhou, Chendi</creatorcontrib><creatorcontrib>Zhou, Jingbo</creatorcontrib><creatorcontrib>Zhou, Yang</creatorcontrib><creatorcontrib>Dai, Huaiyu</creatorcontrib><creatorcontrib>Dou, Dejing</creatorcontrib><title>Multi-Job Intelligent Scheduling with Cross-Device Federated Learning</title><description>Recent years have witnessed a large amount of decentralized data in various (edge) devices of end-users, while the decentralized data aggregation remains complicated for machine learning jobs because of regulations and laws. As a practical approach to handling decentralized data, Federated Learning (FL) enables collaborative global machine learning model training without sharing sensitive raw data. The servers schedule devices to jobs within the training process of FL. In contrast, device scheduling with multiple jobs in FL remains a critical and open problem. In this paper, we propose a novel multi-job FL framework, which enables the training process of multiple jobs in parallel. The multi-job FL framework is composed of a system model and a scheduling method. The system model enables a parallel training process of multiple jobs, with a cost model based on the data fairness and the training time of diverse devices during the parallel training process. We propose a novel intelligent scheduling approach based on multiple scheduling methods, including an original reinforcement learning-based scheduling method and an original Bayesian optimization-based scheduling method, which corresponds to a small cost while scheduling devices to multiple jobs. We conduct extensive experimentation with diverse jobs and datasets. The experimental results reveal that our proposed approaches significantly outperform baseline approaches in terms of training time (up to 12.73 times faster) and accuracy (up to 46.4% higher).</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Distributed, Parallel, and Cluster Computing</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz71OwzAYhWEvDKhwAUz4Bhz8FzsZUWihKIiB7tFn-0tryaTIcQvcPVA6neXVkR5CbgSvdFPX_A7yVzxWUgpRCaUVvyTLl0MqkT3vHV1PBVOKW5wKffM7DIcUpy39jGVHu7yfZ_aAx-iRrjBghoKB9gh5-o2uyMUIacbr8y7IZrXcdE-sf31cd_c9A2M5ay1IzWs9NlZ5GxrTco6utb4B742WXKKSYHxwAF65UdsahTGjko4b64NakNv_25Nj-MjxHfL38OcZTh71A3PURWg</recordid><startdate>20221124</startdate><enddate>20221124</enddate><creator>Liu, Ji</creator><creator>Jia, Juncheng</creator><creator>Ma, Beichen</creator><creator>Zhou, Chendi</creator><creator>Zhou, Jingbo</creator><creator>Zhou, Yang</creator><creator>Dai, Huaiyu</creator><creator>Dou, Dejing</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20221124</creationdate><title>Multi-Job Intelligent Scheduling with Cross-Device Federated Learning</title><author>Liu, Ji ; Jia, Juncheng ; Ma, Beichen ; Zhou, Chendi ; Zhou, Jingbo ; Zhou, Yang ; Dai, Huaiyu ; Dou, Dejing</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-97a24054f873c7d86900eb97c8acc64202e32a6cdbaac3bf475e166f32b067cd3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Distributed, Parallel, and Cluster Computing</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Liu, Ji</creatorcontrib><creatorcontrib>Jia, Juncheng</creatorcontrib><creatorcontrib>Ma, Beichen</creatorcontrib><creatorcontrib>Zhou, Chendi</creatorcontrib><creatorcontrib>Zhou, Jingbo</creatorcontrib><creatorcontrib>Zhou, Yang</creatorcontrib><creatorcontrib>Dai, Huaiyu</creatorcontrib><creatorcontrib>Dou, Dejing</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Liu, Ji</au><au>Jia, Juncheng</au><au>Ma, Beichen</au><au>Zhou, Chendi</au><au>Zhou, Jingbo</au><au>Zhou, Yang</au><au>Dai, Huaiyu</au><au>Dou, Dejing</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multi-Job Intelligent Scheduling with Cross-Device Federated Learning</atitle><date>2022-11-24</date><risdate>2022</risdate><abstract>Recent years have witnessed a large amount of decentralized data in various (edge) devices of end-users, while the decentralized data aggregation remains complicated for machine learning jobs because of regulations and laws. As a practical approach to handling decentralized data, Federated Learning (FL) enables collaborative global machine learning model training without sharing sensitive raw data. The servers schedule devices to jobs within the training process of FL. In contrast, device scheduling with multiple jobs in FL remains a critical and open problem. In this paper, we propose a novel multi-job FL framework, which enables the training process of multiple jobs in parallel. The multi-job FL framework is composed of a system model and a scheduling method. The system model enables a parallel training process of multiple jobs, with a cost model based on the data fairness and the training time of diverse devices during the parallel training process. We propose a novel intelligent scheduling approach based on multiple scheduling methods, including an original reinforcement learning-based scheduling method and an original Bayesian optimization-based scheduling method, which corresponds to a small cost while scheduling devices to multiple jobs. We conduct extensive experimentation with diverse jobs and datasets. The experimental results reveal that our proposed approaches significantly outperform baseline approaches in terms of training time (up to 12.73 times faster) and accuracy (up to 46.4% higher).</abstract><doi>10.48550/arxiv.2211.13430</doi><oa>free_for_read</oa></addata></record>
fulltext fulltext_linktorsrc
identifier DOI: 10.48550/arxiv.2211.13430
ispartof
issn
language eng
recordid cdi_arxiv_primary_2211_13430
source arXiv.org
subjects Computer Science - Artificial Intelligence
Computer Science - Distributed, Parallel, and Cluster Computing
Computer Science - Learning
title Multi-Job Intelligent Scheduling with Cross-Device Federated Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T19%3A02%3A41IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multi-Job%20Intelligent%20Scheduling%20with%20Cross-Device%20Federated%20Learning&rft.au=Liu,%20Ji&rft.date=2022-11-24&rft_id=info:doi/10.48550/arxiv.2211.13430&rft_dat=%3Carxiv_GOX%3E2211_13430%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true