Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks
As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues f...
Gespeichert in:
Veröffentlicht in: | IEEE internet of things journal 2023-02, Vol.10 (3), p.2667-2679 |
---|---|
Hauptverfasser: | , , , , , , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
container_end_page | 2679 |
---|---|
container_issue | 3 |
container_start_page | 2667 |
container_title | IEEE internet of things journal |
container_volume | 10 |
creator | Zhang, Hengxi Lu, Chengyue Tang, Huaze Wei, Xiaoli Liang, Le Cheng, Ling Ding, Wenbo Han, Zhu |
description | As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues for vehicular networks. Nevertheless, due to the strong nonconvexity and nonlinearity, the classical joint resource allocation problem in vehicular networks is typically NP-hard. The multiagent reinforcement learning (MARL) has emerged as a promising solution to tackle this challenge but its stability and scalability are not satisfactory when the amount of vehicles gets increased. In this article, we mainly investigate the issue of joint spectrum and power allocation in vehicular communication networks, and carefully consider the interactions between the vehicles and environment by incorporating the cooperative stochastic game theory with MARL, named complete-game MARL (CG-MARL), to achieve a better convergence and stability with the theoretical computational complexity [Formula Omitted] with [Formula Omitted] denoting the dimension of action space and [Formula Omitted] denoting the number of V2X Vehicular. Furthermore, the mean-field game (MFG) theory is employed to further enhance the MARL for decreasing the horrible computing resource consumption caused by the CG-MARL to [Formula Omitted] while maintaining an approximate performance. The simulation results demonstrate that the proposed mean-field-aided MARL (MF-MARL) for vehicular network resource allocation can achieve 95% near-optimal performance with much lower complexity, which indicates its significant potentials in the scenarios with massive and dense vehicles. |
doi_str_mv | 10.1109/JIOT.2022.3214525 |
format | Article |
fullrecord | <record><control><sourceid>proquest_cross</sourceid><recordid>TN_cdi_proquest_journals_2769391221</recordid><sourceformat>XML</sourceformat><sourcesystem>PC</sourcesystem><sourcerecordid>2769391221</sourcerecordid><originalsourceid>FETCH-LOGICAL-c2243-583dc1a7e53b003e0490ac409dc6696289b0b46c6f7cbb84a9827c8889a7f2063</originalsourceid><addsrcrecordid>eNpNkF1PwjAUhhujiQT5Ad418XrYnm7dekmIKAYkMeht7boOi6PDdovx39sFLrw6H-97PvIgdEvJlFIi7p-Xm-0UCMCUAU0zyC7QCBjkSco5XP7Lr9EkhD0hJI5lVPAR-lgb5ZKFNU2VzGxlKrzum86qnXEdfjXW1a3X5jBUK6O8s26HYytKoe2jgmdN02rV2dZh6_C7-bS6b5THL6b7af1XuEFXtWqCmZzjGL0tHrbzp2S1eVzOZ6tEA6QsyQpWaapyk7GSEGZIKojSKRGV5lxwKERJypRrXue6LItUiQJyXRSFUHkNhLMxujvtPfr2uzehk_v4oIsnJeRcMEEBaHTRk0v7NgRvann09qD8r6REDizlwFIOLOWZJfsD62Fmaw</addsrcrecordid><sourcetype>Aggregation Database</sourcetype><iscdi>true</iscdi><recordtype>article</recordtype><pqid>2769391221</pqid></control><display><type>article</type><title>Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks</title><source>IEEE Electronic Library (IEL)</source><creator>Zhang, Hengxi ; Lu, Chengyue ; Tang, Huaze ; Wei, Xiaoli ; Liang, Le ; Cheng, Ling ; Ding, Wenbo ; Han, Zhu</creator><creatorcontrib>Zhang, Hengxi ; Lu, Chengyue ; Tang, Huaze ; Wei, Xiaoli ; Liang, Le ; Cheng, Ling ; Ding, Wenbo ; Han, Zhu</creatorcontrib><description>As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues for vehicular networks. Nevertheless, due to the strong nonconvexity and nonlinearity, the classical joint resource allocation problem in vehicular networks is typically NP-hard. The multiagent reinforcement learning (MARL) has emerged as a promising solution to tackle this challenge but its stability and scalability are not satisfactory when the amount of vehicles gets increased. In this article, we mainly investigate the issue of joint spectrum and power allocation in vehicular communication networks, and carefully consider the interactions between the vehicles and environment by incorporating the cooperative stochastic game theory with MARL, named complete-game MARL (CG-MARL), to achieve a better convergence and stability with the theoretical computational complexity [Formula Omitted] with [Formula Omitted] denoting the dimension of action space and [Formula Omitted] denoting the number of V2X Vehicular. Furthermore, the mean-field game (MFG) theory is employed to further enhance the MARL for decreasing the horrible computing resource consumption caused by the CG-MARL to [Formula Omitted] while maintaining an approximate performance. The simulation results demonstrate that the proposed mean-field-aided MARL (MF-MARL) for vehicular network resource allocation can achieve 95% near-optimal performance with much lower complexity, which indicates its significant potentials in the scenarios with massive and dense vehicles.</description><identifier>ISSN: 2327-4662</identifier><identifier>EISSN: 2327-4662</identifier><identifier>DOI: 10.1109/JIOT.2022.3214525</identifier><language>eng</language><publisher>Piscataway: The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</publisher><subject>Communication networks ; Complexity ; Game theory ; Machine learning ; Multiagent systems ; Resource allocation ; Stability ; Vehicles</subject><ispartof>IEEE internet of things journal, 2023-02, Vol.10 (3), p.2667-2679</ispartof><rights>Copyright The Institute of Electrical and Electronics Engineers, Inc. (IEEE) 2023</rights><woscitedreferencessubscribed>false</woscitedreferencessubscribed><citedby>FETCH-LOGICAL-c2243-583dc1a7e53b003e0490ac409dc6696289b0b46c6f7cbb84a9827c8889a7f2063</citedby><cites>FETCH-LOGICAL-c2243-583dc1a7e53b003e0490ac409dc6696289b0b46c6f7cbb84a9827c8889a7f2063</cites><orcidid>0000-0002-6606-5822 ; 0000-0002-8489-1933 ; 0000-0002-4463-1766 ; 0000-0001-7873-8206 ; 0000-0002-0597-4512</orcidid></display><links><openurl>$$Topenurl_article</openurl><openurlfulltext>$$Topenurlfull_article</openurlfulltext><thumbnail>$$Tsyndetics_thumb_exl</thumbnail><link.rule.ids>314,777,781,27905,27906</link.rule.ids></links><search><creatorcontrib>Zhang, Hengxi</creatorcontrib><creatorcontrib>Lu, Chengyue</creatorcontrib><creatorcontrib>Tang, Huaze</creatorcontrib><creatorcontrib>Wei, Xiaoli</creatorcontrib><creatorcontrib>Liang, Le</creatorcontrib><creatorcontrib>Cheng, Ling</creatorcontrib><creatorcontrib>Ding, Wenbo</creatorcontrib><creatorcontrib>Han, Zhu</creatorcontrib><title>Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks</title><title>IEEE internet of things journal</title><description>As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues for vehicular networks. Nevertheless, due to the strong nonconvexity and nonlinearity, the classical joint resource allocation problem in vehicular networks is typically NP-hard. The multiagent reinforcement learning (MARL) has emerged as a promising solution to tackle this challenge but its stability and scalability are not satisfactory when the amount of vehicles gets increased. In this article, we mainly investigate the issue of joint spectrum and power allocation in vehicular communication networks, and carefully consider the interactions between the vehicles and environment by incorporating the cooperative stochastic game theory with MARL, named complete-game MARL (CG-MARL), to achieve a better convergence and stability with the theoretical computational complexity [Formula Omitted] with [Formula Omitted] denoting the dimension of action space and [Formula Omitted] denoting the number of V2X Vehicular. Furthermore, the mean-field game (MFG) theory is employed to further enhance the MARL for decreasing the horrible computing resource consumption caused by the CG-MARL to [Formula Omitted] while maintaining an approximate performance. The simulation results demonstrate that the proposed mean-field-aided MARL (MF-MARL) for vehicular network resource allocation can achieve 95% near-optimal performance with much lower complexity, which indicates its significant potentials in the scenarios with massive and dense vehicles.</description><subject>Communication networks</subject><subject>Complexity</subject><subject>Game theory</subject><subject>Machine learning</subject><subject>Multiagent systems</subject><subject>Resource allocation</subject><subject>Stability</subject><subject>Vehicles</subject><issn>2327-4662</issn><issn>2327-4662</issn><fulltext>true</fulltext><rsrctype>article</rsrctype><creationdate>2023</creationdate><recordtype>article</recordtype><recordid>eNpNkF1PwjAUhhujiQT5Ad418XrYnm7dekmIKAYkMeht7boOi6PDdovx39sFLrw6H-97PvIgdEvJlFIi7p-Xm-0UCMCUAU0zyC7QCBjkSco5XP7Lr9EkhD0hJI5lVPAR-lgb5ZKFNU2VzGxlKrzum86qnXEdfjXW1a3X5jBUK6O8s26HYytKoe2jgmdN02rV2dZh6_C7-bS6b5THL6b7af1XuEFXtWqCmZzjGL0tHrbzp2S1eVzOZ6tEA6QsyQpWaapyk7GSEGZIKojSKRGV5lxwKERJypRrXue6LItUiQJyXRSFUHkNhLMxujvtPfr2uzehk_v4oIsnJeRcMEEBaHTRk0v7NgRvann09qD8r6REDizlwFIOLOWZJfsD62Fmaw</recordid><startdate>20230201</startdate><enddate>20230201</enddate><creator>Zhang, Hengxi</creator><creator>Lu, Chengyue</creator><creator>Tang, Huaze</creator><creator>Wei, Xiaoli</creator><creator>Liang, Le</creator><creator>Cheng, Ling</creator><creator>Ding, Wenbo</creator><creator>Han, Zhu</creator><general>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</general><scope>AAYXX</scope><scope>CITATION</scope><scope>7SC</scope><scope>8FD</scope><scope>JQ2</scope><scope>L7M</scope><scope>L~C</scope><scope>L~D</scope><orcidid>https://orcid.org/0000-0002-6606-5822</orcidid><orcidid>https://orcid.org/0000-0002-8489-1933</orcidid><orcidid>https://orcid.org/0000-0002-4463-1766</orcidid><orcidid>https://orcid.org/0000-0001-7873-8206</orcidid><orcidid>https://orcid.org/0000-0002-0597-4512</orcidid></search><sort><creationdate>20230201</creationdate><title>Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks</title><author>Zhang, Hengxi ; Lu, Chengyue ; Tang, Huaze ; Wei, Xiaoli ; Liang, Le ; Cheng, Ling ; Ding, Wenbo ; Han, Zhu</author></sort><facets><frbrtype>5</frbrtype><frbrgroupid>cdi_FETCH-LOGICAL-c2243-583dc1a7e53b003e0490ac409dc6696289b0b46c6f7cbb84a9827c8889a7f2063</frbrgroupid><rsrctype>articles</rsrctype><prefilter>articles</prefilter><language>eng</language><creationdate>2023</creationdate><topic>Communication networks</topic><topic>Complexity</topic><topic>Game theory</topic><topic>Machine learning</topic><topic>Multiagent systems</topic><topic>Resource allocation</topic><topic>Stability</topic><topic>Vehicles</topic><toplevel>online_resources</toplevel><creatorcontrib>Zhang, Hengxi</creatorcontrib><creatorcontrib>Lu, Chengyue</creatorcontrib><creatorcontrib>Tang, Huaze</creatorcontrib><creatorcontrib>Wei, Xiaoli</creatorcontrib><creatorcontrib>Liang, Le</creatorcontrib><creatorcontrib>Cheng, Ling</creatorcontrib><creatorcontrib>Ding, Wenbo</creatorcontrib><creatorcontrib>Han, Zhu</creatorcontrib><collection>CrossRef</collection><collection>Computer and Information Systems Abstracts</collection><collection>Technology Research Database</collection><collection>ProQuest Computer Science Collection</collection><collection>Advanced Technologies Database with Aerospace</collection><collection>Computer and Information Systems Abstracts Academic</collection><collection>Computer and Information Systems Abstracts Professional</collection><jtitle>IEEE internet of things journal</jtitle></facets><delivery><delcategory>Remote Search Resource</delcategory><fulltext>fulltext</fulltext></delivery><addata><au>Zhang, Hengxi</au><au>Lu, Chengyue</au><au>Tang, Huaze</au><au>Wei, Xiaoli</au><au>Liang, Le</au><au>Cheng, Ling</au><au>Ding, Wenbo</au><au>Han, Zhu</au><format>journal</format><genre>article</genre><ristype>JOUR</ristype><atitle>Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks</atitle><jtitle>IEEE internet of things journal</jtitle><date>2023-02-01</date><risdate>2023</risdate><volume>10</volume><issue>3</issue><spage>2667</spage><epage>2679</epage><pages>2667-2679</pages><issn>2327-4662</issn><eissn>2327-4662</eissn><abstract>As one technique for autonomous driving, vehicular networks can achieve high efficiency with vehicle-and-infrastructure cooperation, bringing high safety and many value-added services. To achieve higher communication efficiency, much effort has been done to cope with the resource allocation issues for vehicular networks. Nevertheless, due to the strong nonconvexity and nonlinearity, the classical joint resource allocation problem in vehicular networks is typically NP-hard. The multiagent reinforcement learning (MARL) has emerged as a promising solution to tackle this challenge but its stability and scalability are not satisfactory when the amount of vehicles gets increased. In this article, we mainly investigate the issue of joint spectrum and power allocation in vehicular communication networks, and carefully consider the interactions between the vehicles and environment by incorporating the cooperative stochastic game theory with MARL, named complete-game MARL (CG-MARL), to achieve a better convergence and stability with the theoretical computational complexity [Formula Omitted] with [Formula Omitted] denoting the dimension of action space and [Formula Omitted] denoting the number of V2X Vehicular. Furthermore, the mean-field game (MFG) theory is employed to further enhance the MARL for decreasing the horrible computing resource consumption caused by the CG-MARL to [Formula Omitted] while maintaining an approximate performance. The simulation results demonstrate that the proposed mean-field-aided MARL (MF-MARL) for vehicular network resource allocation can achieve 95% near-optimal performance with much lower complexity, which indicates its significant potentials in the scenarios with massive and dense vehicles.</abstract><cop>Piscataway</cop><pub>The Institute of Electrical and Electronics Engineers, Inc. (IEEE)</pub><doi>10.1109/JIOT.2022.3214525</doi><tpages>13</tpages><orcidid>https://orcid.org/0000-0002-6606-5822</orcidid><orcidid>https://orcid.org/0000-0002-8489-1933</orcidid><orcidid>https://orcid.org/0000-0002-4463-1766</orcidid><orcidid>https://orcid.org/0000-0001-7873-8206</orcidid><orcidid>https://orcid.org/0000-0002-0597-4512</orcidid></addata></record> |
fulltext | fulltext |
identifier | ISSN: 2327-4662 |
ispartof | IEEE internet of things journal, 2023-02, Vol.10 (3), p.2667-2679 |
issn | 2327-4662 2327-4662 |
language | eng |
recordid | cdi_proquest_journals_2769391221 |
source | IEEE Electronic Library (IEL) |
subjects | Communication networks Complexity Game theory Machine learning Multiagent systems Resource allocation Stability Vehicles |
title | Mean-Field-Aided Multiagent Reinforcement Learning for Resource Allocation in Vehicular Networks |
url | https://sfx.bib-bvb.de/sfx_tum?ctx_ver=Z39.88-2004&ctx_enc=info:ofi/enc:UTF-8&ctx_tim=2025-01-18T07%3A00%3A01IST&url_ver=Z39.88-2004&url_ctx_fmt=infofi/fmt:kev:mtx:ctx&rfr_id=info:sid/primo.exlibrisgroup.com:primo3-Article-proquest_cross&rft_val_fmt=info:ofi/fmt:kev:mtx:journal&rft.genre=article&rft.atitle=Mean-Field-Aided%20Multiagent%20Reinforcement%20Learning%20for%20Resource%20Allocation%20in%20Vehicular%20Networks&rft.jtitle=IEEE%20internet%20of%20things%20journal&rft.au=Zhang,%20Hengxi&rft.date=2023-02-01&rft.volume=10&rft.issue=3&rft.spage=2667&rft.epage=2679&rft.pages=2667-2679&rft.issn=2327-4662&rft.eissn=2327-4662&rft_id=info:doi/10.1109/JIOT.2022.3214525&rft_dat=%3Cproquest_cross%3E2769391221%3C/proquest_cross%3E%3Curl%3E%3C/url%3E&disable_directlink=true&sfx.directlink=off&sfx.report_link=0&rft_id=info:oai/&rft_pqid=2769391221&rft_id=info:pmid/&rfr_iscdi=true |