Multi-Job Intelligent Scheduling with Cross-Device Federated Learning
Recent years have witnessed a large amount of decentralized data in various (edge) devices of end-users, while the decentralized data aggregation remains complicated for machine learning jobs because of regulations and laws. As a practical approach to handling decentralized data, Federated Learning...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | |
---|---|
container_issue | |
container_start_page | |
container_title | |
container_volume | |
creator | Liu, Ji Jia, Juncheng Ma, Beichen Zhou, Chendi Zhou, Jingbo Zhou, Yang Dai, Huaiyu Dou, Dejing |
description | Recent years have witnessed a large amount of decentralized data in various
(edge) devices of end-users, while the decentralized data aggregation remains
complicated for machine learning jobs because of regulations and laws. As a
practical approach to handling decentralized data, Federated Learning (FL)
enables collaborative global machine learning model training without sharing
sensitive raw data. The servers schedule devices to jobs within the training
process of FL. In contrast, device scheduling with multiple jobs in FL remains
a critical and open problem. In this paper, we propose a novel multi-job FL
framework, which enables the training process of multiple jobs in parallel. The
multi-job FL framework is composed of a system model and a scheduling method.
The system model enables a parallel training process of multiple jobs, with a
cost model based on the data fairness and the training time of diverse devices
during the parallel training process. We propose a novel intelligent scheduling
approach based on multiple scheduling methods, including an original
reinforcement learning-based scheduling method and an original Bayesian
optimization-based scheduling method, which corresponds to a small cost while
scheduling devices to multiple jobs. We conduct extensive experimentation with
diverse jobs and datasets. The experimental results reveal that our proposed
approaches significantly outperform baseline approaches in terms of training
time (up to 12.73 times faster) and accuracy (up to 46.4% higher). |
doi_str_mv | 10.48550/arxiv.2211.13430 |
format | Article |
fullrecord | <record><control><sourceid>arxiv_GOX</sourceid><recordid>TN_cdi_arxiv_primary_2211_13430</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2211_13430</sourcerecordid><originalsourceid>FETCH-LOGICAL-a670-97a24054f873c7d86900eb97c8acc64202e32a6cdbaac3bf475e166f32b067cd3</originalsourceid><addsrcrecordid>eNotz71OwzAYhWEvDKhwAUz4Bhz8FzsZUWihKIiB7tFn-0tryaTIcQvcPVA6neXVkR5CbgSvdFPX_A7yVzxWUgpRCaUVvyTLl0MqkT3vHV1PBVOKW5wKffM7DIcUpy39jGVHu7yfZ_aAx-iRrjBghoKB9gh5-o2uyMUIacbr8y7IZrXcdE-sf31cd_c9A2M5ay1IzWs9NlZ5GxrTco6utb4B742WXKKSYHxwAF65UdsahTGjko4b64NakNv_25Nj-MjxHfL38OcZTh71A3PURWg</addsrcrecordid><sourcetype>Open Access Repository</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Multi-Job Intelligent Scheduling with Cross-Device Federated Learning</title><source>arXiv.org</source><creator>Liu, Ji ; Jia, Juncheng ; Ma, Beichen ; Zhou, Chendi ; Zhou, Jingbo ; Zhou, Yang ; Dai, Huaiyu ; Dou, Dejing</creator><creatorcontrib>Liu, Ji ; Jia, Juncheng ; Ma, Beichen ; Zhou, Chendi ; Zhou, Jingbo ; Zhou, Yang ; Dai, Huaiyu ; Dou, Dejing</creatorcontrib><description>Recent years have witnessed a large amount of decentralized data in various
(edge) devices of end-users, while the decentralized data aggregation remains
complicated for machine learning jobs because of regulations and laws. As a
practical approach to handling decentralized data, Federated Learning (FL)
enables collaborative global machine learning model training without sharing
sensitive raw data. The servers schedule devices to jobs within the training
process of FL. In contrast, device scheduling with multiple jobs in FL remains
a critical and open problem. In this paper, we propose a novel multi-job FL
framework, which enables the training process of multiple jobs in parallel. The
multi-job FL framework is composed of a system model and a scheduling method.
The system model enables a parallel training process of multiple jobs, with a
cost model based on the data fairness and the training time of diverse devices
during the parallel training process. We propose a novel intelligent scheduling
approach based on multiple scheduling methods, including an original
reinforcement learning-based scheduling method and an original Bayesian
optimization-based scheduling method, which corresponds to a small cost while
scheduling devices to multiple jobs. We conduct extensive experimentation with
diverse jobs and datasets. The experimental results reveal that our proposed
approaches significantly outperform baseline approaches in terms of training
time (up to 12.73 times faster) and accuracy (up to 46.4% higher).</description><identifier>DOI: 10.48550/arxiv.2211.13430</identifier><language>eng</language><subject>Computer Science - Artificial Intelligence ; Computer Science - Distributed, Parallel, and Cluster Computing ; Computer Science - Learning</subject><creationdate>2022-11</creationdate><rights>http://creativecommons.org/licenses/by/4.0</rights><oa>free_for_read</oa><woscitedreferencessubscribed>false</woscitedreferencessubscribed></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>228,230,780,885</link.rule.ids><linktorsrc>$$Uhttps://arxiv.org/abs/2211.13430$$EView_record_in_Cornell_University$$FView_record_in_$$GCornell_University$$Hfree_for_read</linktorsrc><backlink>$$Uhttps://doi.org/10.48550/arXiv.2211.13430$$DView paper in arXiv$$Hfree_for_read</backlink></links><search><creatorcontrib>Liu, Ji</creatorcontrib><creatorcontrib>Jia, Juncheng</creatorcontrib><creatorcontrib>Ma, Beichen</creatorcontrib><creatorcontrib>Zhou, Chendi</creatorcontrib><creatorcontrib>Zhou, Jingbo</creatorcontrib><creatorcontrib>Zhou, Yang</creatorcontrib><creatorcontrib>Dai, Huaiyu</creatorcontrib><creatorcontrib>Dou, Dejing</creatorcontrib><title>Multi-Job Intelligent Scheduling with Cross-Device Federated Learning</title><description>Recent years have witnessed a large amount of decentralized data in various
(edge) devices of end-users, while the decentralized data aggregation remains
complicated for machine learning jobs because of regulations and laws. As a
practical approach to handling decentralized data, Federated Learning (FL)
enables collaborative global machine learning model training without sharing
sensitive raw data. The servers schedule devices to jobs within the training
process of FL. In contrast, device scheduling with multiple jobs in FL remains
a critical and open problem. In this paper, we propose a novel multi-job FL
framework, which enables the training process of multiple jobs in parallel. The
multi-job FL framework is composed of a system model and a scheduling method.
The system model enables a parallel training process of multiple jobs, with a
cost model based on the data fairness and the training time of diverse devices
during the parallel training process. We propose a novel intelligent scheduling
approach based on multiple scheduling methods, including an original
reinforcement learning-based scheduling method and an original Bayesian
optimization-based scheduling method, which corresponds to a small cost while
scheduling devices to multiple jobs. We conduct extensive experimentation with
diverse jobs and datasets. The experimental results reveal that our proposed
approaches significantly outperform baseline approaches in terms of training
time (up to 12.73 times faster) and accuracy (up to 46.4% higher).</description><subject>Computer Science - Artificial Intelligence</subject><subject>Computer Science - Distributed, Parallel, and Cluster Computing</subject><subject>Computer Science - Learning</subject><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2022</creationdate><recordtype>article</recordtype><sourceid>GOX</sourceid><recordid>eNotz71OwzAYhWEvDKhwAUz4Bhz8FzsZUWihKIiB7tFn-0tryaTIcQvcPVA6neXVkR5CbgSvdFPX_A7yVzxWUgpRCaUVvyTLl0MqkT3vHV1PBVOKW5wKffM7DIcUpy39jGVHu7yfZ_aAx-iRrjBghoKB9gh5-o2uyMUIacbr8y7IZrXcdE-sf31cd_c9A2M5ay1IzWs9NlZ5GxrTco6utb4B742WXKKSYHxwAF65UdsahTGjko4b64NakNv_25Nj-MjxHfL38OcZTh71A3PURWg</recordid><startdate>20221124</startdate><enddate>20221124</enddate><creator>Liu, Ji</creator><creator>Jia, Juncheng</creator><creator>Ma, Beichen</creator><creator>Zhou, Chendi</creator><creator>Zhou, Jingbo</creator><creator>Zhou, Yang</creator><creator>Dai, Huaiyu</creator><creator>Dou, Dejing</creator><scope>AKY</scope><scope>GOX</scope></search><sort><creationdate>20221124</creationdate><title>Multi-Job Intelligent Scheduling with Cross-Device Federated Learning</title><author>Liu, Ji ; Jia, Juncheng ; Ma, Beichen ; Zhou, Chendi ; Zhou, Jingbo ; Zhou, Yang ; Dai, Huaiyu ; Dou, Dejing</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-a670-97a24054f873c7d86900eb97c8acc64202e32a6cdbaac3bf475e166f32b067cd3</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2022</creationdate><topic>Computer Science - Artificial Intelligence</topic><topic>Computer Science - Distributed, Parallel, and Cluster Computing</topic><topic>Computer Science - Learning</topic><toplevel>online_resources</toplevel><creatorcontrib>Liu, Ji</creatorcontrib><creatorcontrib>Jia, Juncheng</creatorcontrib><creatorcontrib>Ma, Beichen</creatorcontrib><creatorcontrib>Zhou, Chendi</creatorcontrib><creatorcontrib>Zhou, Jingbo</creatorcontrib><creatorcontrib>Zhou, Yang</creatorcontrib><creatorcontrib>Dai, Huaiyu</creatorcontrib><creatorcontrib>Dou, Dejing</creatorcontrib><collection>arXiv Computer Science</collection><collection>arXiv.org</collection></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Liu, Ji</au><au>Jia, Juncheng</au><au>Ma, Beichen</au><au>Zhou, Chendi</au><au>Zhou, Jingbo</au><au>Zhou, Yang</au><au>Dai, Huaiyu</au><au>Dou, Dejing</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Multi-Job Intelligent Scheduling with Cross-Device Federated Learning</atitle><date>2022-11-24</date><risdate>2022</risdate><abstract>Recent years have witnessed a large amount of decentralized data in various
(edge) devices of end-users, while the decentralized data aggregation remains
complicated for machine learning jobs because of regulations and laws. As a
practical approach to handling decentralized data, Federated Learning (FL)
enables collaborative global machine learning model training without sharing
sensitive raw data. The servers schedule devices to jobs within the training
process of FL. In contrast, device scheduling with multiple jobs in FL remains
a critical and open problem. In this paper, we propose a novel multi-job FL
framework, which enables the training process of multiple jobs in parallel. The
multi-job FL framework is composed of a system model and a scheduling method.
The system model enables a parallel training process of multiple jobs, with a
cost model based on the data fairness and the training time of diverse devices
during the parallel training process. We propose a novel intelligent scheduling
approach based on multiple scheduling methods, including an original
reinforcement learning-based scheduling method and an original Bayesian
optimization-based scheduling method, which corresponds to a small cost while
scheduling devices to multiple jobs. We conduct extensive experimentation with
diverse jobs and datasets. The experimental results reveal that our proposed
approaches significantly outperform baseline approaches in terms of training
time (up to 12.73 times faster) and accuracy (up to 46.4% higher).</abstract><doi>10.48550/arxiv.2211.13430</doi><oa>free_for_read</oa></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | DOI: 10.48550/arxiv.2211.13430 |
ispartof | |
issn | |
language | eng |
recordid | cdi_arxiv_primary_2211_13430 |
source | arXiv.org |
subjects | Computer Science - Artificial Intelligence Computer Science - Distributed, Parallel, and Cluster Computing Computer Science - Learning |
title | Multi-Job Intelligent Scheduling with Cross-Device Federated Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-03T19%3A02%3A41IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-arxiv_GOX&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Multi-Job%20Intelligent%20Scheduling%20with%20Cross-Device%20Federated%20Learning&rft.au=Liu,%20Ji&rft.date=2022-11-24&rft_id=info:doi/10.48550/arxiv.2211.13430&rft_dat=%3Carxiv_GOX%3E2211_13430%3C/arxiv_GOX%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rfr_iscdi=true |