Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning
Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training t...
Gespeichert in:
Veröffentlicht in: | IEEE eTransactions on network and service management 2024-08, p.1-1 |
---|---|
Hauptverfasser: | , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 1 |
---|---|
container_issue | |
container_start_page | 1 |
container_title | IEEE eTransactions on network and service management |
container_volume | |
creator | Zhou, Cong Zhao, Baokang Tang, Fengxiao Han, Biao Wang, Baosheng |
description | Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training time are too long. Concomitantly, they are unable to satisfy when the number of objectives is dynamic. It is necessary to model service function chain placement as a multi-objective problem. The multi-objective problem can decomposed into multiple subproblems by the weight vectors. In this paper, we first reveal the relationship between weight vectors and solution position, which can reduce the training time to gain a better placement model. Then, we design a novel algorithm for the service function chain placement problem, called rzMODRL. The weight vectors are divided into zones for training in parallel, and the order is defined for the final models located at the end of a training process, which can save time and improve the quality of the model. Dynamic objective placement method is based on the high-dimensional model to avoid retraining for a low-dimensional placement. Evaluation results show that the proposed algorithms improve the service acceptance ratio up to 32% and the hyper-volume values with 14% in the multi-objective service function chain placement, where hyper-volume has been widely applied to evaluate the convergence and diversity simultaneously in multi-objective optimization. And it is also effective in solving the dynamic objective service function chain placement problem that the difference of average hyper-volume values is 10.44% |
doi_str_mv | 10.1109/TNSM.2024.3446248 |
format | Article |
fullrecord | <record><control><sourceid>ieee_RIE</sourceid><recordid>TN_cdi_ieee_primary_10640148</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><ieee_id>10640148</ieee_id><sourcerecordid>10640148</sourcerecordid><originalsourceid>FETCH-ieee_primary_106401483</originalsourceid><addsrcrecordid>eNqFyc0KgkAUQOEhCLKfBwhazAtoM-Noui2LFhmRbkMmu9YVHWPUoLevRftWB85HyJwzh3MWLtNjEjuCCem4UvpCBgNi8dAVtvTc1YiM27ZkzAt4KCxyid5a1ZjTuK86tJtrCXmHL6AJmBfmQHe9_o5G081DoaanSuVQg-7oWrVwo1-IAJ70DKiLxvzsAMpo1PcpGRaqamH264Qsdtt0s7cRALKnwVqZd8aZLxmXgfuHPzUJQsg</addsrcrecordid><sourcetype>Publisher</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype></control><display><type>article</type><title>Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning</title><source>IEEE Electronic Library (IEL)</source><creator>Zhou, Cong ; Zhao, Baokang ; Tang, Fengxiao ; Han, Biao ; Wang, Baosheng</creator><creatorcontrib>Zhou, Cong ; Zhao, Baokang ; Tang, Fengxiao ; Han, Biao ; Wang, Baosheng</creatorcontrib><description>Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training time are too long. Concomitantly, they are unable to satisfy when the number of objectives is dynamic. It is necessary to model service function chain placement as a multi-objective problem. The multi-objective problem can decomposed into multiple subproblems by the weight vectors. In this paper, we first reveal the relationship between weight vectors and solution position, which can reduce the training time to gain a better placement model. Then, we design a novel algorithm for the service function chain placement problem, called rzMODRL. The weight vectors are divided into zones for training in parallel, and the order is defined for the final models located at the end of a training process, which can save time and improve the quality of the model. Dynamic objective placement method is based on the high-dimensional model to avoid retraining for a low-dimensional placement. Evaluation results show that the proposed algorithms improve the service acceptance ratio up to 32% and the hyper-volume values with 14% in the multi-objective service function chain placement, where hyper-volume has been widely applied to evaluate the convergence and diversity simultaneously in multi-objective optimization. And it is also effective in solving the dynamic objective service function chain placement problem that the difference of average hyper-volume values is 10.44%</description><identifier>EISSN: 1932-4537</identifier><identifier>DOI: 10.1109/TNSM.2024.3446248</identifier><identifier>CODEN: ITNSC4</identifier><language>eng</language><publisher>IEEE</publisher><subject>Computational modeling ; Deep reinforcement learning ; Heuristic algorithms ; multi-objective deep reinforcement learning ; Optimization ; related zone decomposition ; service function chain placement ; Service function chaining ; Training ; Vectors</subject><ispartof>IEEE eTransactions on network and service management, 2024-08, p.1-1</ispartof><lds50>peer_reviewed</lds50><woscitedreferencessubscribed>false</woscitedreferencessubscribed><orcidid>0000-0002-5082-5727 ; 0000-0002-4703-3577 ; 0000-0001-9200-9018 ; 0000-0003-2414-4802</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><linktohtml>$$Uhttps://ieeexplore.ieee.org/document/10640148$$EHTML$$P50$$Gieee$$H</linktohtml><link.rule.ids>314,780,784,796,27924,27925,54758</link.rule.ids><linktorsrc>$$Uhttps://ieeexplore.ieee.org/document/10640148$$EView_record_in_IEEE$$FView_record_in_$$GIEEE</linktorsrc></links><search><creatorcontrib>Zhou, Cong</creatorcontrib><creatorcontrib>Zhao, Baokang</creatorcontrib><creatorcontrib>Tang, Fengxiao</creatorcontrib><creatorcontrib>Han, Biao</creatorcontrib><creatorcontrib>Wang, Baosheng</creatorcontrib><title>Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning</title><title>IEEE eTransactions on network and service management</title><addtitle>T-NSM</addtitle><description>Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training time are too long. Concomitantly, they are unable to satisfy when the number of objectives is dynamic. It is necessary to model service function chain placement as a multi-objective problem. The multi-objective problem can decomposed into multiple subproblems by the weight vectors. In this paper, we first reveal the relationship between weight vectors and solution position, which can reduce the training time to gain a better placement model. Then, we design a novel algorithm for the service function chain placement problem, called rzMODRL. The weight vectors are divided into zones for training in parallel, and the order is defined for the final models located at the end of a training process, which can save time and improve the quality of the model. Dynamic objective placement method is based on the high-dimensional model to avoid retraining for a low-dimensional placement. Evaluation results show that the proposed algorithms improve the service acceptance ratio up to 32% and the hyper-volume values with 14% in the multi-objective service function chain placement, where hyper-volume has been widely applied to evaluate the convergence and diversity simultaneously in multi-objective optimization. And it is also effective in solving the dynamic objective service function chain placement problem that the difference of average hyper-volume values is 10.44%</description><subject>Computational modeling</subject><subject>Deep reinforcement learning</subject><subject>Heuristic algorithms</subject><subject>multi-objective deep reinforcement learning</subject><subject>Optimization</subject><subject>related zone decomposition</subject><subject>service function chain placement</subject><subject>Service function chaining</subject><subject>Training</subject><subject>Vectors</subject><issn>1932-4537</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2024</creationdate><recordtype>article</recordtype><sourceid>RIE</sourceid><recordid>eNqFyc0KgkAUQOEhCLKfBwhazAtoM-Noui2LFhmRbkMmu9YVHWPUoLevRftWB85HyJwzh3MWLtNjEjuCCem4UvpCBgNi8dAVtvTc1YiM27ZkzAt4KCxyid5a1ZjTuK86tJtrCXmHL6AJmBfmQHe9_o5G081DoaanSuVQg-7oWrVwo1-IAJ70DKiLxvzsAMpo1PcpGRaqamH264Qsdtt0s7cRALKnwVqZd8aZLxmXgfuHPzUJQsg</recordid><startdate>20240819</startdate><enddate>20240819</enddate><creator>Zhou, Cong</creator><creator>Zhao, Baokang</creator><creator>Tang, Fengxiao</creator><creator>Han, Biao</creator><creator>Wang, Baosheng</creator><general>IEEE</general><scope>97E</scope><scope>RIA</scope><scope>RIE</scope><orcidid>https://orcid.org/0000-0002-5082-5727</orcidid><orcidid>https://orcid.org/0000-0002-4703-3577</orcidid><orcidid>https://orcid.org/0000-0001-9200-9018</orcidid><orcidid>https://orcid.org/0000-0003-2414-4802</orcidid></search><sort><creationdate>20240819</creationdate><title>Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning</title><author>Zhou, Cong ; Zhao, Baokang ; Tang, Fengxiao ; Han, Biao ; Wang, Baosheng</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-ieee_primary_106401483</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2024</creationdate><topic>Computational modeling</topic><topic>Deep reinforcement learning</topic><topic>Heuristic algorithms</topic><topic>multi-objective deep reinforcement learning</topic><topic>Optimization</topic><topic>related zone decomposition</topic><topic>service function chain placement</topic><topic>Service function chaining</topic><topic>Training</topic><topic>Vectors</topic><toplevel>peer_reviewed</toplevel><toplevel>online_resources</toplevel><creatorcontrib>Zhou, Cong</creatorcontrib><creatorcontrib>Zhao, Baokang</creatorcontrib><creatorcontrib>Tang, Fengxiao</creatorcontrib><creatorcontrib>Han, Biao</creatorcontrib><creatorcontrib>Wang, Baosheng</creatorcontrib><collection>IEEE All-Society Periodicals Package (ASPP) 2005-present</collection><collection>IEEE All-Society Periodicals Package (ASPP) 1998-Present</collection><collection>IEEE Electronic Library (IEL)</collection><jtitle>IEEE eTransactions on network and service management</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext_linktorsrc</fulltext></delivery><addata><au>Zhou, Cong</au><au>Zhao, Baokang</au><au>Tang, Fengxiao</au><au>Han, Biao</au><au>Wang, Baosheng</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning</atitle><jtitle>IEEE eTransactions on network and service management</jtitle><stitle>T-NSM</stitle><date>2024-08-19</date><risdate>2024</risdate><spage>1</spage><epage>1</epage><pages>1-1</pages><eissn>1932-4537</eissn><coden>ITNSC4</coden><abstract>Service function chain placement is crucial to support services flexibility and diversity for different users and vendors. Specifically, this problem is proved to be NP-hard. Existing deep reinforcement learning based methods either can only handle a limited number of objectives, or their training time are too long. Concomitantly, they are unable to satisfy when the number of objectives is dynamic. It is necessary to model service function chain placement as a multi-objective problem. The multi-objective problem can decomposed into multiple subproblems by the weight vectors. In this paper, we first reveal the relationship between weight vectors and solution position, which can reduce the training time to gain a better placement model. Then, we design a novel algorithm for the service function chain placement problem, called rzMODRL. The weight vectors are divided into zones for training in parallel, and the order is defined for the final models located at the end of a training process, which can save time and improve the quality of the model. Dynamic objective placement method is based on the high-dimensional model to avoid retraining for a low-dimensional placement. Evaluation results show that the proposed algorithms improve the service acceptance ratio up to 32% and the hyper-volume values with 14% in the multi-objective service function chain placement, where hyper-volume has been widely applied to evaluate the convergence and diversity simultaneously in multi-objective optimization. And it is also effective in solving the dynamic objective service function chain placement problem that the difference of average hyper-volume values is 10.44%</abstract><pub>IEEE</pub><doi>10.1109/TNSM.2024.3446248</doi><orcidid>https://orcid.org/0000-0002-5082-5727</orcidid><orcidid>https://orcid.org/0000-0002-4703-3577</orcidid><orcidid>https://orcid.org/0000-0001-9200-9018</orcidid><orcidid>https://orcid.org/0000-0003-2414-4802</orcidid></addata></record> |
fulltext | fulltext_linktorsrc |
identifier | EISSN: 1932-4537 |
ispartof | IEEE eTransactions on network and service management, 2024-08, p.1-1 |
issn | 1932-4537 |
language | eng |
recordid | cdi_ieee_primary_10640148 |
source | IEEE Electronic Library (IEL) |
subjects | Computational modeling Deep reinforcement learning Heuristic algorithms multi-objective deep reinforcement learning Optimization related zone decomposition service function chain placement Service function chaining Training Vectors |
title | Dynamic Multi-objective Service Function Chain Placement Based on Deep Reinforcement Learning |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2024-12-29T13%3A06%3A45IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-ieee_RIE&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Dynamic%20Multi-objective%20Service%20Function%20Chain%20Placement%20Based%20on%20Deep%20Reinforcement%20Learning&rft.jtitle=IEEE%20eTransactions%20on%20network%20and%20service%20management&rft.au=Zhou,%20Cong&rft.date=2024-08-19&rft.spage=1&rft.epage=1&rft.pages=1-1&rft.eissn=1932-4537&rft.coden=ITNSC4&rft_id=info:doi/10.1109/TNSM.2024.3446248&rft_dat=%3Cieee_RIE%3E10640148%3C/ieee_RIE%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_id=info:pmid/&rft_ieee_id=10640148&rfr_iscdi=true |