Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning

Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training t...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE eTransactions on network and service management 2024-08, p.1-1
Hauptverfasser: Zhou, Cong, Zhao, Baokang, Tang, Fengxiao, Han, Biao, Wang, Baosheng
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext bestellen
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
container_end_page 1
container_issue
container_start_page 1
container_title IEEE eTransactions on network and service management
container_volume
creator Zhou, Cong
Zhao, Baokang
Tang, Fengxiao
Han, Biao
Wang, Baosheng
description Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training time are too long. Concomitantly, they are unable to satisfy when the number of objectives is dynamic. It is necessary to model service function chain placement as a multi-objective problem. The multi-objective problem can decomposed into multiple subproblems by the weight vectors. In this paper, we first reveal the relationship between weight vectors and solution position, which can reduce the training time to gain a better placement model. Then, we design a novel algorithm for the service function chain placement problem, called rzMODRL. The weight vectors are divided into zones for training in parallel, and the order is defined for the final models located at the end of a training process, which can save time and improve the quality of the model. Dynamic objective placement method is based on the high-dimensional model to avoid retraining for a low-dimensional placement. Evaluation results show that the proposed algorithms improve the service acceptance ratio up to 32% and the hyper-volume values with 14% in the multi-objective service function chain placement, where hyper-volume has been widely applied to evaluate the convergence and diversity simultaneously in multi-objective optimization. And it is also effective in solving the dynamic objective service function chain placement problem that the difference of average hyper-volume values is 10.44%
doi_str_mv 10.1109/TNSM.2024.3446248
format Article
fullrecord <record><control><sourceid>ieee_RIE</sourceid><recordid>TN_cdi_ieee_primary_10640148</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10640148</ieee_id><sourcerecordid>10640148</sourcerecordid><originalsourceid>FETCH-ieee_primary_106401483</originalsourceid><addsrcrecordid>eNqFyc0KgkAUQOEhCLKfBwhazAtoM-Noui2LFhmRbkMmu9YVHWPUoLevRftWB85HyJwzh3MWLtNjEjuCCem4UvpCBgNi8dAVtvTc1YiM27ZkzAt4KCxyid5a1ZjTuK86tJtrCXmHL6AJmBfmQHe9_o5G081DoaanSuVQg-7oWrVwo1-IAJ70DKiLxvzsAMpo1PcpGRaqamH264Qsdtt0s7cRALKnwVqZd8aZLxmXgfuHPzUJQsg</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning</title><source>IEEE Electronic Library (IEL)</source><creator>Zhou, Cong ; Zhao, Baokang ; Tang, Fengxiao ; Han, Biao ; Wang, Baosheng</creator><creatorcontrib>Zhou, Cong ; Zhao, Baokang ; Tang, Fengxiao ; Han, Biao ; Wang, Baosheng</creatorcontrib><description>Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training time are too long. Concomitantly, they are unable to satisfy when the number of objectives is dynamic. It is necessary to model service function chain placement as a multi-objective problem. The multi-objective problem can decomposed into multiple subproblems by the weight vectors. In this paper, we first reveal the relationship between weight vectors and solution position, which can reduce the training time to gain a better placement model. Then, we design a novel algorithm for the service function chain placement problem, called rzMODRL. The weight vectors are divided into zones for training in parallel, and the order is defined for the final models located at the end of a training process, which can save time and improve the quality of the model. Dynamic objective placement method is based on the high-dimensional model to avoid retraining for a low-dimensional placement. Evaluation results show that the proposed algorithms improve the service acceptance ratio up to 32% and the hyper-volume values with 14% in the multi-objective service function chain placement, where hyper-volume has been widely applied to evaluate the convergence and diversity simultaneously in multi-objective optimization. And it is also effective in solving the dynamic objective service function chain placement problem that the difference of average hyper-volume values is 10.44%</description><identifier>EISSN: 1932-4537</identifier><identifier>DOI: 10.1109/TNSM.2024.3446248</identifier><identifier>CODEN: ITNSC4</identifier><language>eng</language><publisher>IEEE</publisher><subject>Computational modeling ; Deep reinforcement learning ; Heuristic algorithms ; multi-objective deep reinforcement learning ; Optimization ; related zone decomposition ; service function chain placement ; Service function chaining ; Training ; Vectors</subject><ispartof>IEEE eTransactions on network and service management, 2024-08, p.1-1</ispartof><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>0000-0002-5082-5727 ; 0000-0002-4703-3577 ; 0000-0001-9200-9018 ; 0000-0003-2414-4802</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10640148$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,796,27924,27925,54758</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10640148$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Zhou, Cong</creatorcontrib><creatorcontrib>Zhao, Baokang</creatorcontrib><creatorcontrib>Tang, Fengxiao</creatorcontrib><creatorcontrib>Han, Biao</creatorcontrib><creatorcontrib>Wang, Baosheng</creatorcontrib><title>Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning</title><title>IEEE eTransactions on network and service management</title><addtitle>T-NSM</addtitle><description>Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training time are too long. Concomitantly, they are unable to satisfy when the number of objectives is dynamic. It is necessary to model service function chain placement as a multi-objective problem. The multi-objective problem can decomposed into multiple subproblems by the weight vectors. In this paper, we first reveal the relationship between weight vectors and solution position, which can reduce the training time to gain a better placement model. Then, we design a novel algorithm for the service function chain placement problem, called rzMODRL. The weight vectors are divided into zones for training in parallel, and the order is defined for the final models located at the end of a training process, which can save time and improve the quality of the model. Dynamic objective placement method is based on the high-dimensional model to avoid retraining for a low-dimensional placement. Evaluation results show that the proposed algorithms improve the service acceptance ratio up to 32% and the hyper-volume values with 14% in the multi-objective service function chain placement, where hyper-volume has been widely applied to evaluate the convergence and diversity simultaneously in multi-objective optimization. And it is also effective in solving the dynamic objective service function chain placement problem that the difference of average hyper-volume values is 10.44%</description><subject>Computational modeling</subject><subject>Deep reinforcement learning</subject><subject>Heuristic algorithms</subject><subject>multi-objective deep reinforcement learning</subject><subject>Optimization</subject><subject>related zone decomposition</subject><subject>service function chain placement</subject><subject>Service function chaining</subject><subject>Training</subject><subject>Vectors</subject><issn>1932-4537</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNqFyc0KgkAUQOEhCLKfBwhazAtoM-Noui2LFhmRbkMmu9YVHWPUoLevRftWB85HyJwzh3MWLtNjEjuCCem4UvpCBgNi8dAVtvTc1YiM27ZkzAt4KCxyid5a1ZjTuK86tJtrCXmHL6AJmBfmQHe9_o5G081DoaanSuVQg-7oWrVwo1-IAJ70DKiLxvzsAMpo1PcpGRaqamH264Qsdtt0s7cRALKnwVqZd8aZLxmXgfuHPzUJQsg</recordid><startdate>20240819</startdate><enddate>20240819</enddate><creator>Zhou, Cong</creator><creator>Zhao, Baokang</creator><creator>Tang, Fengxiao</creator><creator>Han, Biao</creator><creator>Wang, Baosheng</creator><general>IEEE</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><orcidid>https://orcid.org/0000-0002-5082-5727</orcidid><orcidid>https://orcid.org/0000-0002-4703-3577</orcidid><orcidid>https://orcid.org/0000-0001-9200-9018</orcidid><orcidid>https://orcid.org/0000-0003-2414-4802</orcidid></search><sort><creationdate>20240819</creationdate><title>Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning</title><author>Zhou, Cong ; Zhao, Baokang ; Tang, Fengxiao ; Han, Biao ; Wang, Baosheng</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-ieee_primary_106401483</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computational modeling</topic><topic>Deep reinforcement learning</topic><topic>Heuristic algorithms</topic><topic>multi-objective deep reinforcement learning</topic><topic>Optimization</topic><topic>related zone decomposition</topic><topic>service function chain placement</topic><topic>Service function chaining</topic><topic>Training</topic><topic>Vectors</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhou, Cong</creatorcontrib><creatorcontrib>Zhao, Baokang</creatorcontrib><creatorcontrib>Tang, Fengxiao</creatorcontrib><creatorcontrib>Han, Biao</creatorcontrib><creatorcontrib>Wang, Baosheng</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><jtitle>IEEE eTransactions on network and service management</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zhou, Cong</au><au>Zhao, Baokang</au><au>Tang, Fengxiao</au><au>Han, Biao</au><au>Wang, Baosheng</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning</atitle><jtitle>IEEE eTransactions on network and service management</jtitle><stitle>T-NSM</stitle><date>2024-08-19</date><risdate>2024</risdate><spage>1</spage><epage>1</epage><pages>1-1</pages><eissn>1932-4537</eissn><coden>ITNSC4</coden><abstract>Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training time are too long. Concomitantly, they are unable to satisfy when the number of objectives is dynamic. It is necessary to model service function chain placement as a multi-objective problem. The multi-objective problem can decomposed into multiple subproblems by the weight vectors. In this paper, we first reveal the relationship between weight vectors and solution position, which can reduce the training time to gain a better placement model. Then, we design a novel algorithm for the service function chain placement problem, called rzMODRL. The weight vectors are divided into zones for training in parallel, and the order is defined for the final models located at the end of a training process, which can save time and improve the quality of the model. Dynamic objective placement method is based on the high-dimensional model to avoid retraining for a low-dimensional placement. Evaluation results show that the proposed algorithms improve the service acceptance ratio up to 32% and the hyper-volume values with 14% in the multi-objective service function chain placement, where hyper-volume has been widely applied to evaluate the convergence and diversity simultaneously in multi-objective optimization. And it is also effective in solving the dynamic objective service function chain placement problem that the difference of average hyper-volume values is 10.44%</abstract><pub>IEEE</pub><doi>10.1109/TNSM.2024.3446248</doi><orcidid>https://orcid.org/0000-0002-5082-5727</orcidid><orcidid>https://orcid.org/0000-0002-4703-3577</orcidid><orcidid>https://orcid.org/0000-0001-9200-9018</orcidid><orcidid>https://orcid.org/0000-0003-2414-4802</orcidid></addata></record>
fulltext fulltext_linktorsrc
identifier EISSN: 1932-4537
ispartof IEEE eTransactions on network and service management, 2024-08, p.1-1
issn 1932-4537
language eng
recordid cdi_ieee_primary_10640148
source IEEE Electronic Library (IEL)
subjects Computational modeling
Deep reinforcement learning
Heuristic algorithms
multi-objective deep reinforcement learning
Optimization
related zone decomposition
service function chain placement
Service function chaining
Training
Vectors
title Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning
url https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-29T13%3A06%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Dynamic%20Multi-objective%20Service%20Function%20Chain%20Placement%20Based%20on%20Deep%20Reinforcement%20Learning&rft.jtitle=IEEE%20eTransactions%20on%20network%20and%20service%20management&rft.au=Zhou,%20Cong&rft.date=2024-08-19&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.eissn=1932-4537&rft.coden=ITNSC4&rft_id=info:doi/10.1109/TNSM.2024.3446248&rft_dat=%3Cieee_RIE%3E10640148%3C/ieee_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10640148&rfr_iscdi=true